var/home/core/zuul-output/0000755000175000017500000000000015114721600014522 5ustar corecorevar/home/core/zuul-output/logs/0000755000175000017500000000000015114726366015504 5ustar corecorevar/home/core/zuul-output/logs/kubelet.log0000644000000000000000004104373015114726357017712 0ustar rootrootDec 06 03:33:59 crc systemd[1]: Starting Kubernetes Kubelet... Dec 06 03:33:59 crc restorecon[4679]: Relabeled /var/lib/kubelet/config.json from system_u:object_r:unlabeled_t:s0 to system_u:object_r:container_var_lib_t:s0 Dec 06 03:33:59 crc restorecon[4679]: /var/lib/kubelet/device-plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 06 03:33:59 crc restorecon[4679]: /var/lib/kubelet/device-plugins/kubelet.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 06 03:33:59 crc restorecon[4679]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/volumes/kubernetes.io~configmap/nginx-conf/..2025_02_23_05_40_35.4114275528/nginx.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Dec 06 03:33:59 crc restorecon[4679]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Dec 06 03:33:59 crc restorecon[4679]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/22e96971 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Dec 06 03:33:59 crc restorecon[4679]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/21c98286 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Dec 06 03:33:59 crc restorecon[4679]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/0f1869e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Dec 06 03:33:59 crc restorecon[4679]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Dec 06 03:33:59 crc restorecon[4679]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/46889d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Dec 06 03:33:59 crc restorecon[4679]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/5b6a5969 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Dec 06 03:33:59 crc restorecon[4679]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/6c7921f5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Dec 06 03:33:59 crc restorecon[4679]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4804f443 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Dec 06 03:33:59 crc restorecon[4679]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/2a46b283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Dec 06 03:33:59 crc restorecon[4679]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/a6b5573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Dec 06 03:33:59 crc restorecon[4679]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4f88ee5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Dec 06 03:33:59 crc restorecon[4679]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/5a4eee4b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Dec 06 03:33:59 crc restorecon[4679]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/cd87c521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Dec 06 03:33:59 crc restorecon[4679]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 06 03:33:59 crc restorecon[4679]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 06 03:33:59 crc restorecon[4679]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 06 03:33:59 crc restorecon[4679]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 06 03:33:59 crc restorecon[4679]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 06 03:33:59 crc restorecon[4679]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 06 03:33:59 crc restorecon[4679]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/38602af4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 06 03:33:59 crc restorecon[4679]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/1483b002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 06 03:33:59 crc restorecon[4679]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/0346718b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 06 03:33:59 crc restorecon[4679]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/d3ed4ada not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 06 03:33:59 crc restorecon[4679]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/3bb473a5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 06 03:33:59 crc restorecon[4679]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/8cd075a9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 06 03:33:59 crc restorecon[4679]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/00ab4760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 06 03:33:59 crc restorecon[4679]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/54a21c09 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 06 03:33:59 crc restorecon[4679]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Dec 06 03:33:59 crc restorecon[4679]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/70478888 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Dec 06 03:33:59 crc restorecon[4679]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/43802770 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Dec 06 03:33:59 crc restorecon[4679]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/955a0edc not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Dec 06 03:33:59 crc restorecon[4679]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/bca2d009 not reset as customized by admin to system_u:object_r:container_file_t:s0:c140,c1009 Dec 06 03:33:59 crc restorecon[4679]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/b295f9bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Dec 06 03:33:59 crc restorecon[4679]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 06 03:33:59 crc restorecon[4679]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 06 03:33:59 crc restorecon[4679]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 06 03:33:59 crc restorecon[4679]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 06 03:33:59 crc restorecon[4679]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 06 03:33:59 crc restorecon[4679]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 06 03:33:59 crc restorecon[4679]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 06 03:33:59 crc restorecon[4679]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 06 03:33:59 crc restorecon[4679]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 06 03:33:59 crc restorecon[4679]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 06 03:33:59 crc restorecon[4679]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 06 03:33:59 crc restorecon[4679]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/bc46ea27 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 06 03:33:59 crc restorecon[4679]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5731fc1b not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 06 03:33:59 crc restorecon[4679]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5e1b2a3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 06 03:33:59 crc restorecon[4679]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/943f0936 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 06 03:33:59 crc restorecon[4679]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/3f764ee4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 06 03:33:59 crc restorecon[4679]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/8695e3f9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 06 03:33:59 crc restorecon[4679]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/aed7aa86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 06 03:33:59 crc restorecon[4679]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/c64d7448 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 06 03:33:59 crc restorecon[4679]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/0ba16bd2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 06 03:33:59 crc restorecon[4679]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/207a939f not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 06 03:33:59 crc restorecon[4679]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/54aa8cdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 06 03:33:59 crc restorecon[4679]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/1f5fa595 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 06 03:33:59 crc restorecon[4679]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/bf9c8153 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 06 03:33:59 crc restorecon[4679]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/47fba4ea not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 06 03:33:59 crc restorecon[4679]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/7ae55ce9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 06 03:33:59 crc restorecon[4679]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7906a268 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 06 03:33:59 crc restorecon[4679]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/ce43fa69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 06 03:33:59 crc restorecon[4679]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7fc7ea3a not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 06 03:33:59 crc restorecon[4679]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/d8c38b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 06 03:33:59 crc restorecon[4679]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/9ef015fb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 06 03:33:59 crc restorecon[4679]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/b9db6a41 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 06 03:33:59 crc restorecon[4679]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Dec 06 03:33:59 crc restorecon[4679]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/b1733d79 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Dec 06 03:33:59 crc restorecon[4679]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/afccd338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Dec 06 03:33:59 crc restorecon[4679]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/9df0a185 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Dec 06 03:33:59 crc restorecon[4679]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/18938cf8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Dec 06 03:33:59 crc restorecon[4679]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/7ab4eb23 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Dec 06 03:33:59 crc restorecon[4679]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/56930be6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Dec 06 03:33:59 crc restorecon[4679]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 06 03:33:59 crc restorecon[4679]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_35.630010865 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 06 03:33:59 crc restorecon[4679]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 06 03:33:59 crc restorecon[4679]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 06 03:33:59 crc restorecon[4679]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 06 03:33:59 crc restorecon[4679]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 06 03:33:59 crc restorecon[4679]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 06 03:33:59 crc restorecon[4679]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 06 03:33:59 crc restorecon[4679]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 06 03:33:59 crc restorecon[4679]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/0d8e3722 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Dec 06 03:33:59 crc restorecon[4679]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/d22b2e76 not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Dec 06 03:33:59 crc restorecon[4679]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/e036759f not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 06 03:33:59 crc restorecon[4679]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/2734c483 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Dec 06 03:33:59 crc restorecon[4679]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/57878fe7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Dec 06 03:33:59 crc restorecon[4679]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/3f3c2e58 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Dec 06 03:33:59 crc restorecon[4679]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/375bec3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Dec 06 03:33:59 crc restorecon[4679]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/7bc41e08 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 06 03:33:59 crc restorecon[4679]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 06 03:33:59 crc restorecon[4679]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/48c7a72d not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 06 03:33:59 crc restorecon[4679]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/4b66701f not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 06 03:33:59 crc restorecon[4679]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/a5a1c202 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 06 03:33:59 crc restorecon[4679]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 06 03:33:59 crc restorecon[4679]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 06 03:33:59 crc restorecon[4679]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 06 03:33:59 crc restorecon[4679]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 06 03:33:59 crc restorecon[4679]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 06 03:33:59 crc restorecon[4679]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 06 03:33:59 crc restorecon[4679]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 06 03:33:59 crc restorecon[4679]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 06 03:33:59 crc restorecon[4679]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_40.1388695756 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 06 03:33:59 crc restorecon[4679]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 06 03:33:59 crc restorecon[4679]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 06 03:33:59 crc restorecon[4679]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/26f3df5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 06 03:33:59 crc restorecon[4679]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/6d8fb21d not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 06 03:33:59 crc restorecon[4679]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/50e94777 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 06 03:33:59 crc restorecon[4679]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208473b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 06 03:33:59 crc restorecon[4679]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/ec9e08ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 06 03:33:59 crc restorecon[4679]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3b787c39 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 06 03:33:59 crc restorecon[4679]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208eaed5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 06 03:33:59 crc restorecon[4679]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/93aa3a2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 06 03:33:59 crc restorecon[4679]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3c697968 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 06 03:33:59 crc restorecon[4679]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 06 03:33:59 crc restorecon[4679]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/ba950ec9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 06 03:33:59 crc restorecon[4679]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/cb5cdb37 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 06 03:33:59 crc restorecon[4679]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/f2df9827 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 06 03:33:59 crc restorecon[4679]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 06 03:33:59 crc restorecon[4679]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 06 03:33:59 crc restorecon[4679]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 06 03:33:59 crc restorecon[4679]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 06 03:33:59 crc restorecon[4679]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 06 03:33:59 crc restorecon[4679]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 06 03:33:59 crc restorecon[4679]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 06 03:33:59 crc restorecon[4679]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 06 03:33:59 crc restorecon[4679]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 06 03:33:59 crc restorecon[4679]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 06 03:33:59 crc restorecon[4679]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 06 03:33:59 crc restorecon[4679]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/fedaa673 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 06 03:33:59 crc restorecon[4679]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/9ca2df95 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 06 03:33:59 crc restorecon[4679]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/b2d7460e not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 06 03:33:59 crc restorecon[4679]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2207853c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 06 03:33:59 crc restorecon[4679]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/241c1c29 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 06 03:33:59 crc restorecon[4679]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2d910eaf not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 06 03:33:59 crc restorecon[4679]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 06 03:33:59 crc restorecon[4679]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 06 03:33:59 crc restorecon[4679]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 06 03:33:59 crc restorecon[4679]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 06 03:33:59 crc restorecon[4679]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 06 03:33:59 crc restorecon[4679]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 06 03:33:59 crc restorecon[4679]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 06 03:33:59 crc restorecon[4679]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 06 03:33:59 crc restorecon[4679]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 06 03:33:59 crc restorecon[4679]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 06 03:33:59 crc restorecon[4679]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 06 03:33:59 crc restorecon[4679]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 06 03:33:59 crc restorecon[4679]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 06 03:33:59 crc restorecon[4679]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 06 03:33:59 crc restorecon[4679]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 06 03:33:59 crc restorecon[4679]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 06 03:33:59 crc restorecon[4679]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/c6c0f2e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Dec 06 03:33:59 crc restorecon[4679]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/399edc97 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Dec 06 03:33:59 crc restorecon[4679]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8049f7cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Dec 06 03:33:59 crc restorecon[4679]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/0cec5484 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Dec 06 03:33:59 crc restorecon[4679]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/312446d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c406,c828 Dec 06 03:33:59 crc restorecon[4679]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8e56a35d not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 06 03:33:59 crc restorecon[4679]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 06 03:33:59 crc restorecon[4679]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 06 03:33:59 crc restorecon[4679]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 06 03:33:59 crc restorecon[4679]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 06 03:33:59 crc restorecon[4679]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 06 03:33:59 crc restorecon[4679]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 06 03:33:59 crc restorecon[4679]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/2d30ddb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Dec 06 03:33:59 crc restorecon[4679]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/eca8053d not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Dec 06 03:33:59 crc restorecon[4679]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/c3a25c9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c168,c522 Dec 06 03:33:59 crc restorecon[4679]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/b9609c22 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 06 03:33:59 crc restorecon[4679]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Dec 06 03:33:59 crc restorecon[4679]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/e8b0eca9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Dec 06 03:33:59 crc restorecon[4679]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/b36a9c3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Dec 06 03:33:59 crc restorecon[4679]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/38af7b07 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Dec 06 03:33:59 crc restorecon[4679]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/ae821620 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Dec 06 03:33:59 crc restorecon[4679]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/baa23338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Dec 06 03:33:59 crc restorecon[4679]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/2c534809 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Dec 06 03:33:59 crc restorecon[4679]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 06 03:33:59 crc restorecon[4679]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 06 03:33:59 crc restorecon[4679]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 06 03:33:59 crc restorecon[4679]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 06 03:33:59 crc restorecon[4679]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 06 03:33:59 crc restorecon[4679]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 06 03:33:59 crc restorecon[4679]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/59b29eae not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Dec 06 03:33:59 crc restorecon[4679]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/c91a8e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Dec 06 03:33:59 crc restorecon[4679]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/4d87494a not reset as customized by admin to system_u:object_r:container_file_t:s0:c442,c857 Dec 06 03:33:59 crc restorecon[4679]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/1e33ca63 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 06 03:33:59 crc restorecon[4679]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 06 03:33:59 crc restorecon[4679]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/8dea7be2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 06 03:33:59 crc restorecon[4679]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d0b04a99 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 06 03:33:59 crc restorecon[4679]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d84f01e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 06 03:33:59 crc restorecon[4679]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/4109059b not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 06 03:33:59 crc restorecon[4679]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/a7258a3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 06 03:33:59 crc restorecon[4679]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/05bdf2b6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 06 03:33:59 crc restorecon[4679]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 06 03:33:59 crc restorecon[4679]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/f3261b51 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 06 03:33:59 crc restorecon[4679]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/315d045e not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 06 03:33:59 crc restorecon[4679]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/5fdcf278 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 06 03:33:59 crc restorecon[4679]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/d053f757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 06 03:33:59 crc restorecon[4679]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/c2850dc7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 06 03:33:59 crc restorecon[4679]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:33:59 crc restorecon[4679]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:33:59 crc restorecon[4679]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:33:59 crc restorecon[4679]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:33:59 crc restorecon[4679]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:33:59 crc restorecon[4679]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:33:59 crc restorecon[4679]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fcfb0b2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:33:59 crc restorecon[4679]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c7ac9b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:33:59 crc restorecon[4679]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fa0c0d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:33:59 crc restorecon[4679]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c609b6ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:33:59 crc restorecon[4679]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/2be6c296 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:33:59 crc restorecon[4679]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/89a32653 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:33:59 crc restorecon[4679]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/4eb9afeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:33:59 crc restorecon[4679]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/13af6efa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:33:59 crc restorecon[4679]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 06 03:33:59 crc restorecon[4679]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/b03f9724 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 06 03:33:59 crc restorecon[4679]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/e3d105cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 06 03:33:59 crc restorecon[4679]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/3aed4d83 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 06 03:33:59 crc restorecon[4679]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 06 03:33:59 crc restorecon[4679]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 06 03:33:59 crc restorecon[4679]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 06 03:33:59 crc restorecon[4679]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 06 03:33:59 crc restorecon[4679]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 06 03:33:59 crc restorecon[4679]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 06 03:33:59 crc restorecon[4679]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/0765fa6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 06 03:33:59 crc restorecon[4679]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/2cefc627 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 06 03:33:59 crc restorecon[4679]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/3dcc6345 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 06 03:33:59 crc restorecon[4679]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/365af391 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 06 03:33:59 crc restorecon[4679]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 06 03:33:59 crc restorecon[4679]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 06 03:33:59 crc restorecon[4679]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 06 03:33:59 crc restorecon[4679]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 06 03:33:59 crc restorecon[4679]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 06 03:33:59 crc restorecon[4679]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 06 03:33:59 crc restorecon[4679]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 06 03:33:59 crc restorecon[4679]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 06 03:33:59 crc restorecon[4679]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b1130c0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 06 03:33:59 crc restorecon[4679]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/236a5913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b9432e26 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/5ddb0e3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/986dc4fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/8a23ff9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/9728ae68 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/665f31d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/136c9b42 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/98a1575b not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/cac69136 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/5deb77a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/2ae53400 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/e46f2326 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/dc688d3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/3497c3cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/177eb008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/af5a2afa not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/d780cb1f not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/49b0f374 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/26fbb125 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/cf14125a not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/b7f86972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/e51d739c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/88ba6a69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/669a9acf not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/5cd51231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/75349ec7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/15c26839 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/45023dcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/2bb66a50 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/64d03bdd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/ab8e7ca0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/bb9be25f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/9a0b61d3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/d471b9d2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/8cb76b8e not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/11a00840 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/ec355a92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/992f735e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d59cdbbc not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/72133ff0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/c56c834c not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d13724c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/0a498258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa471982 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fc900d92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa7d68da not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/4bacf9b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/424021b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/fc2e31a3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/f51eefac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/c8997f2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/7481f599 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/fdafea19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/d0e1c571 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/ee398915 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/682bb6b8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a3e67855 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a989f289 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/915431bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/7796fdab not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/dcdb5f19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/a3aaa88c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/5508e3e6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/160585de not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/e99f8da3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/8bc85570 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/a5861c91 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/84db1135 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/9e1a6043 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/c1aba1c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/d55ccd6d not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/971cc9f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/8f2e3dcf not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/ceb35e9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/1c192745 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/5209e501 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/f83de4df not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/e7b978ac not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/c64304a1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/5384386b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/cce3e3ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/8fb75465 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/740f573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/32fd1134 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/0a861bd3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/80363026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/bfa952a8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..2025_02_23_05_33_31.333075221 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/793bf43d not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/7db1bb6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/4f6a0368 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/c12c7d86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/36c4a773 not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/4c1e98ae not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/a4c8115c not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/setup/7db1802e not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver/a008a7ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-syncer/2c836bac not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-regeneration-controller/0ce62299 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-insecure-readyz/945d2457 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-check-endpoints/7d5c1dd8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/index.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/bundle-v1.15.0.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/channel.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/package.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/bc8d0691 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/6b76097a not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/34d1af30 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/312ba61c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/645d5dd1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/16e825f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/4cf51fc9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/2a23d348 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/075dbd49 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/dd585ddd not reset as customized by admin to system_u:object_r:container_file_t:s0:c377,c642 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/17ebd0ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c343 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/005579f4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_23_11.1287037894 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/bf5f3b9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/af276eb7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/ea28e322 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/692e6683 not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/871746a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/4eb2e958 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/ca9b62da not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/0edd6fce not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/containers/controller-manager/89b4555f not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/655fcd71 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/0d43c002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/e68efd17 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/9acf9b65 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/5ae3ff11 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/1e59206a not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/27af16d1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c304,c1017 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/7918e729 not reset as customized by admin to system_u:object_r:container_file_t:s0:c853,c893 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/5d976d0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c585,c981 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/d7f55cbb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/f0812073 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/1a56cbeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/7fdd437e not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/cdfb5652 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/fix-audit-permissions/fb93119e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver/f1e8fc0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver-check-endpoints/218511f3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server/serving-certs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/ca8af7b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/72cc8a75 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/6e8a3760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4c3455c0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/2278acb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4b453e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/3ec09bda not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2/cacerts.bin not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java/cacerts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl/ca-bundle.trust.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/email-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/objsign-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2ae6433e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fde84897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75680d2e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/openshift-service-serving-signer_1740288168.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/facfc4fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f5a969c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CFCA_EV_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9ef4a08a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ingress-operator_1740288202.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2f332aed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/248c8271.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d10a21f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ACCVRAIZ1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a94d09e5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c9a4d3b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40193066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd8c0d63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b936d1c6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CA_Disig_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4fd49c6c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM_SERVIDORES_SEGUROS.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b81b93f0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f9a69fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b30d5fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ANF_Secure_Server_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b433981b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93851c9e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9282e51c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7dd1bc4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Actalis_Authentication_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/930ac5d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f47b495.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e113c810.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5931b5bc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Commercial.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2b349938.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e48193cf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/302904dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a716d4ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Networking.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93bc0acc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/86212b19.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b727005e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbc54cab.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f51bb24c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c28a8a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9c8dfbd4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ccc52f49.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cb1c3204.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ce5e74ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd08c599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6d41d539.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb5fa911.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e35234b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8cb5ee0f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a7c655d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f8fc53da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/de6d66f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d41b5e2a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/41a3f684.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1df5a75f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_2011.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e36a6752.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b872f2b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9576d26b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/228f89db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_ECC_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb717492.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d21b73c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b1b94ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/595e996b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_RSA_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b46e03d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/128f4b91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_3_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81f2d2b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Autoridad_de_Certificacion_Firmaprofesional_CIF_A62634068.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3bde41ac.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d16a5865.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_EC-384_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0179095f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ffa7f1eb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9482e63a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4dae3dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e359ba6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7e067d03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/95aff9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7746a63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Baltimore_CyberTrust_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/653b494a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3ad48a91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_2_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/54657681.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/82223c44.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8de2f56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d9dafe4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d96b65e2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee64a828.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40547a79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5a3f0ff8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a780d93.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/34d996fb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/eed8c118.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/89c02a45.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b1159c4c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d6325660.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4c339cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8312c4c1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_E1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8508e720.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5fdd185d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48bec511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/69105f4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b9bc432.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/32888f65.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b03dec0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/219d9499.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5acf816d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbf06781.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc99f41e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AAA_Certificate_Services.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/985c1f52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8794b4e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_BR_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7c037b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ef954a4e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_EV_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2add47b6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/90c5a3c8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0f3e76e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/53a1b57a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_EV_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5ad8a5d6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/68dd7389.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d04f354.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d6437c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/062cdee6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bd43e1dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7f3d5d1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c491639e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3513523f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/399e7759.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/feffd413.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d18e9066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/607986c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c90bc37d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1b0f7e5c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e08bfd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dd8e9d41.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed39abd0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a3418fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bc3f2570.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_High_Assurance_EV_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/244b5494.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81b9768f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4be590e0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_ECC_P384_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9846683b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/252252d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e8e7201.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_RSA4096_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d52c538d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c44cc0c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Trusted_Root_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75d1b2ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a2c66da8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ecccd8db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust.net_Certification_Authority__2048_.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/aee5f10d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e7271e8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0e59380.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4c3982f2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b99d060.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf64f35b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0a775a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/002c0b4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cc450945.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_EC1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/106f3e4d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b3fb433b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4042bcee.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/02265526.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/455f1b52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0d69c7e1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9f727ac7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5e98733a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0cd152c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc4d6a89.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6187b673.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/FIRMAPROFESIONAL_CA_ROOT-A_WEB.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ba8887ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/068570d1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f081611a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48a195d8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GDCA_TrustAUTH_R5_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f6fa695.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab59055e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b92fd57f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GLOBALTRUST_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fa5da96b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ec40989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7719f463.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1001acf7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f013ecaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/626dceaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c559d742.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1d3472b9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9479c8c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a81e292b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4bfab552.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e071171e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/57bcb2da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_ECC_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab5346f4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5046c355.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_RSA_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/865fbdf9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da0cfd1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/85cde254.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_ECC_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbb3f32b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureSign_RootCA11.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5860aaa6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/31188b5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HiPKI_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c7f1359b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f15c80c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hongkong_Post_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/09789157.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/18856ac4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e09d511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Commercial_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cf701eeb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d06393bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Public_Sector_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/10531352.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Izenpe.com.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureTrust_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0ed035a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsec_e-Szigno_Root_CA_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8160b96c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8651083.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2c63f966.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_ECC_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d89cda1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/01419da9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_RSA_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7a5b843.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_RSA_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf53fb88.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9591a472.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3afde786.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Gold_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NAVER_Global_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3fb36b73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d39b0a2c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a89d74c2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd58d51e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7db1890.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NetLock_Arany__Class_Gold__F__tan__s__tv__ny.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/988a38cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/60afe812.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f39fc864.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5443e9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GB_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e73d606e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dfc0fe80.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b66938e9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e1eab7c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GC_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/773e07ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c899c73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d59297b8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ddcda989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_1_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/749e9e03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/52b525c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7e8dc79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a819ef2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/08063a00.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b483515.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/064e0aa9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1f58a078.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6f7454b3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7fa05551.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76faf6c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9339512a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f387163d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee37c333.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e18bfb83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e442e424.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fe8a2cd8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/23f4c490.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5cd81ad7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0c70a8d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7892ad52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SZAFIR_ROOT_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4f316efb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_RSA_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/06dc52d5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/583d0756.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0bf05006.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/88950faa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9046744a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c860d51.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_RSA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6fa5da56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/33ee480d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Secure_Global_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/63a2c897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_ECC_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bdacca6f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ff34af3f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbff3a01.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_ECC_RootCA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_C1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/406c9bb1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_C3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Services_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Silver_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/99e1b953.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/14bc7599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TUBITAK_Kamu_SM_SSL_Kok_Sertifikasi_-_Surum_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a3adc42.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f459871d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_ECC_Root_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_RSA_Root_2023.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TeliaSonera_Root_CA_v1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telia_Root_CA_v2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f103249.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f058632f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-certificates.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9bf03295.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/98aaf404.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1cef98f5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/073bfcc5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2923b3f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f249de83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/edcbddb5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P256_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b5697b0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ae85e5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b74d2bd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P384_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d887a5bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9aef356c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TunTrust_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd64f3fc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e13665f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Extended_Validation_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f5dc4f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da7377f6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Global_G2_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c01eb047.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/304d27c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed858448.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f30dd6ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/04f60c28.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_ECC_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fc5a8f99.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/35105088.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee532fd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/XRamp_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/706f604c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76579174.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d86cdd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/882de061.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f618aec.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a9d40e02.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e-Szigno_Root_CA_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e868b802.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/83e9984f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ePKI_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca6e4ad9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d6523ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4b718d9b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/869fbf79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/containers/registry/f8d22bdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/6e8bbfac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/54dd7996 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/a4f1bb05 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/207129da not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/c1df39e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/15b8f1cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/77bd6913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/2382c1b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/704ce128 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/70d16fe0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/bfb95535 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/57a8e8e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/1b9d3e5e not reset as customized by admin to system_u:object_r:container_file_t:s0:c107,c917 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/fddb173c not reset as customized by admin to system_u:object_r:container_file_t:s0:c202,c983 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/95d3c6c4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/bfb5fff5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/2aef40aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/c0391cad not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/1119e69d not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/660608b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/8220bd53 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/85f99d5c not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/4b0225f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/9c2a3394 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/e820b243 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/1ca52ea0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/e6988e45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/6655f00b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/98bc3986 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/08e3458a not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/2a191cb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/6c4eeefb not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/f61a549c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/24891863 not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/fbdfd89c not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/9b63b3bc not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/8acde6d6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/node-driver-registrar/59ecbba3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/csi-provisioner/685d4be3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/containers/route-controller-manager/feaea55e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/63709497 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/d966b7fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/f5773757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/81c9edb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/57bf57ee not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/86f5e6aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/0aabe31d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/d2af85c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/09d157d9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c0fe7256 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c30319e4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/e6b1dd45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/2bb643f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/920de426 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/70fa1e87 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/a1c12a2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/9442e6c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/5b45ec72 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/3c9f3a59 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/1091c11b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/9a6821c6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/ec0c35e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/517f37e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/6214fe78 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/ba189c8b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/351e4f31 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/c0f219ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/8069f607 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/559c3d82 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/605ad488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/148df488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/3bf6dcb4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/022a2feb not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/938c3924 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/729fe23e not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/1fd5cbd4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/a96697e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/e155ddca not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/10dd0e0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/6f2c8392 not reset as customized by admin to system_u:object_r:container_file_t:s0:c267,c588 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/bd241ad9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/plugins/csi-hostpath not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/plugins/csi-hostpath/csi.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/plugins/kubernetes.io not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/plugins/kubernetes.io/csi not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983 not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/vol_data.json not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 06 03:34:00 crc restorecon[4679]: /var/lib/kubelet/plugins_registry not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 06 03:34:00 crc restorecon[4679]: Relabeled /var/usrlocal/bin/kubenswrapper from system_u:object_r:bin_t:s0 to system_u:object_r:kubelet_exec_t:s0 Dec 06 03:34:00 crc kubenswrapper[4980]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Dec 06 03:34:00 crc kubenswrapper[4980]: Flag --minimum-container-ttl-duration has been deprecated, Use --eviction-hard or --eviction-soft instead. Will be removed in a future version. Dec 06 03:34:00 crc kubenswrapper[4980]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Dec 06 03:34:00 crc kubenswrapper[4980]: Flag --register-with-taints has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Dec 06 03:34:00 crc kubenswrapper[4980]: Flag --pod-infra-container-image has been deprecated, will be removed in a future release. Image garbage collector will get sandbox image information from CRI. Dec 06 03:34:00 crc kubenswrapper[4980]: Flag --system-reserved has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Dec 06 03:34:00 crc kubenswrapper[4980]: I1206 03:34:00.949324 4980 server.go:211] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Dec 06 03:34:00 crc kubenswrapper[4980]: W1206 03:34:00.953688 4980 feature_gate.go:330] unrecognized feature gate: NewOLM Dec 06 03:34:00 crc kubenswrapper[4980]: W1206 03:34:00.953731 4980 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Dec 06 03:34:00 crc kubenswrapper[4980]: W1206 03:34:00.953741 4980 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Dec 06 03:34:00 crc kubenswrapper[4980]: W1206 03:34:00.953749 4980 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Dec 06 03:34:00 crc kubenswrapper[4980]: W1206 03:34:00.953759 4980 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Dec 06 03:34:00 crc kubenswrapper[4980]: W1206 03:34:00.953768 4980 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Dec 06 03:34:00 crc kubenswrapper[4980]: W1206 03:34:00.953776 4980 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Dec 06 03:34:00 crc kubenswrapper[4980]: W1206 03:34:00.953785 4980 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Dec 06 03:34:00 crc kubenswrapper[4980]: W1206 03:34:00.953793 4980 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Dec 06 03:34:00 crc kubenswrapper[4980]: W1206 03:34:00.953800 4980 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Dec 06 03:34:00 crc kubenswrapper[4980]: W1206 03:34:00.953808 4980 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Dec 06 03:34:00 crc kubenswrapper[4980]: W1206 03:34:00.953814 4980 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Dec 06 03:34:00 crc kubenswrapper[4980]: W1206 03:34:00.953821 4980 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Dec 06 03:34:00 crc kubenswrapper[4980]: W1206 03:34:00.953829 4980 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Dec 06 03:34:00 crc kubenswrapper[4980]: W1206 03:34:00.953837 4980 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Dec 06 03:34:00 crc kubenswrapper[4980]: W1206 03:34:00.953845 4980 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Dec 06 03:34:00 crc kubenswrapper[4980]: W1206 03:34:00.953852 4980 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Dec 06 03:34:00 crc kubenswrapper[4980]: W1206 03:34:00.953858 4980 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Dec 06 03:34:00 crc kubenswrapper[4980]: W1206 03:34:00.953865 4980 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Dec 06 03:34:00 crc kubenswrapper[4980]: W1206 03:34:00.953872 4980 feature_gate.go:330] unrecognized feature gate: PlatformOperators Dec 06 03:34:00 crc kubenswrapper[4980]: W1206 03:34:00.953878 4980 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Dec 06 03:34:00 crc kubenswrapper[4980]: W1206 03:34:00.953886 4980 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Dec 06 03:34:00 crc kubenswrapper[4980]: W1206 03:34:00.953903 4980 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Dec 06 03:34:00 crc kubenswrapper[4980]: W1206 03:34:00.953910 4980 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Dec 06 03:34:00 crc kubenswrapper[4980]: W1206 03:34:00.953917 4980 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Dec 06 03:34:00 crc kubenswrapper[4980]: W1206 03:34:00.953923 4980 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Dec 06 03:34:00 crc kubenswrapper[4980]: W1206 03:34:00.953929 4980 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Dec 06 03:34:00 crc kubenswrapper[4980]: W1206 03:34:00.953936 4980 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Dec 06 03:34:00 crc kubenswrapper[4980]: W1206 03:34:00.953943 4980 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Dec 06 03:34:00 crc kubenswrapper[4980]: W1206 03:34:00.953951 4980 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Dec 06 03:34:00 crc kubenswrapper[4980]: W1206 03:34:00.953957 4980 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Dec 06 03:34:00 crc kubenswrapper[4980]: W1206 03:34:00.953964 4980 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Dec 06 03:34:00 crc kubenswrapper[4980]: W1206 03:34:00.953970 4980 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Dec 06 03:34:00 crc kubenswrapper[4980]: W1206 03:34:00.953977 4980 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Dec 06 03:34:00 crc kubenswrapper[4980]: W1206 03:34:00.953986 4980 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Dec 06 03:34:00 crc kubenswrapper[4980]: W1206 03:34:00.953995 4980 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Dec 06 03:34:00 crc kubenswrapper[4980]: W1206 03:34:00.954002 4980 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Dec 06 03:34:00 crc kubenswrapper[4980]: W1206 03:34:00.954011 4980 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Dec 06 03:34:00 crc kubenswrapper[4980]: W1206 03:34:00.954021 4980 feature_gate.go:330] unrecognized feature gate: InsightsConfig Dec 06 03:34:00 crc kubenswrapper[4980]: W1206 03:34:00.954029 4980 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Dec 06 03:34:00 crc kubenswrapper[4980]: W1206 03:34:00.954039 4980 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Dec 06 03:34:00 crc kubenswrapper[4980]: W1206 03:34:00.954048 4980 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Dec 06 03:34:00 crc kubenswrapper[4980]: W1206 03:34:00.954055 4980 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Dec 06 03:34:00 crc kubenswrapper[4980]: W1206 03:34:00.954063 4980 feature_gate.go:330] unrecognized feature gate: Example Dec 06 03:34:00 crc kubenswrapper[4980]: W1206 03:34:00.954070 4980 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Dec 06 03:34:00 crc kubenswrapper[4980]: W1206 03:34:00.954077 4980 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Dec 06 03:34:00 crc kubenswrapper[4980]: W1206 03:34:00.954083 4980 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Dec 06 03:34:00 crc kubenswrapper[4980]: W1206 03:34:00.954089 4980 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Dec 06 03:34:00 crc kubenswrapper[4980]: W1206 03:34:00.954096 4980 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Dec 06 03:34:00 crc kubenswrapper[4980]: W1206 03:34:00.954103 4980 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Dec 06 03:34:00 crc kubenswrapper[4980]: W1206 03:34:00.954109 4980 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Dec 06 03:34:00 crc kubenswrapper[4980]: W1206 03:34:00.954115 4980 feature_gate.go:330] unrecognized feature gate: GatewayAPI Dec 06 03:34:00 crc kubenswrapper[4980]: W1206 03:34:00.954122 4980 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Dec 06 03:34:00 crc kubenswrapper[4980]: W1206 03:34:00.954128 4980 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Dec 06 03:34:00 crc kubenswrapper[4980]: W1206 03:34:00.954134 4980 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Dec 06 03:34:00 crc kubenswrapper[4980]: W1206 03:34:00.954143 4980 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Dec 06 03:34:00 crc kubenswrapper[4980]: W1206 03:34:00.954149 4980 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Dec 06 03:34:00 crc kubenswrapper[4980]: W1206 03:34:00.954155 4980 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Dec 06 03:34:00 crc kubenswrapper[4980]: W1206 03:34:00.954162 4980 feature_gate.go:330] unrecognized feature gate: SignatureStores Dec 06 03:34:00 crc kubenswrapper[4980]: W1206 03:34:00.954169 4980 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Dec 06 03:34:00 crc kubenswrapper[4980]: W1206 03:34:00.954174 4980 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Dec 06 03:34:00 crc kubenswrapper[4980]: W1206 03:34:00.954180 4980 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Dec 06 03:34:00 crc kubenswrapper[4980]: W1206 03:34:00.954186 4980 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Dec 06 03:34:00 crc kubenswrapper[4980]: W1206 03:34:00.954192 4980 feature_gate.go:330] unrecognized feature gate: PinnedImages Dec 06 03:34:00 crc kubenswrapper[4980]: W1206 03:34:00.954198 4980 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Dec 06 03:34:00 crc kubenswrapper[4980]: W1206 03:34:00.954205 4980 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Dec 06 03:34:00 crc kubenswrapper[4980]: W1206 03:34:00.954211 4980 feature_gate.go:330] unrecognized feature gate: OVNObservability Dec 06 03:34:00 crc kubenswrapper[4980]: W1206 03:34:00.954217 4980 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Dec 06 03:34:00 crc kubenswrapper[4980]: W1206 03:34:00.954223 4980 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Dec 06 03:34:00 crc kubenswrapper[4980]: W1206 03:34:00.954232 4980 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Dec 06 03:34:00 crc kubenswrapper[4980]: W1206 03:34:00.954238 4980 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Dec 06 03:34:00 crc kubenswrapper[4980]: I1206 03:34:00.954578 4980 flags.go:64] FLAG: --address="0.0.0.0" Dec 06 03:34:00 crc kubenswrapper[4980]: I1206 03:34:00.954603 4980 flags.go:64] FLAG: --allowed-unsafe-sysctls="[]" Dec 06 03:34:00 crc kubenswrapper[4980]: I1206 03:34:00.954615 4980 flags.go:64] FLAG: --anonymous-auth="true" Dec 06 03:34:00 crc kubenswrapper[4980]: I1206 03:34:00.954624 4980 flags.go:64] FLAG: --application-metrics-count-limit="100" Dec 06 03:34:00 crc kubenswrapper[4980]: I1206 03:34:00.954634 4980 flags.go:64] FLAG: --authentication-token-webhook="false" Dec 06 03:34:00 crc kubenswrapper[4980]: I1206 03:34:00.954641 4980 flags.go:64] FLAG: --authentication-token-webhook-cache-ttl="2m0s" Dec 06 03:34:00 crc kubenswrapper[4980]: I1206 03:34:00.954651 4980 flags.go:64] FLAG: --authorization-mode="AlwaysAllow" Dec 06 03:34:00 crc kubenswrapper[4980]: I1206 03:34:00.954660 4980 flags.go:64] FLAG: --authorization-webhook-cache-authorized-ttl="5m0s" Dec 06 03:34:00 crc kubenswrapper[4980]: I1206 03:34:00.954667 4980 flags.go:64] FLAG: --authorization-webhook-cache-unauthorized-ttl="30s" Dec 06 03:34:00 crc kubenswrapper[4980]: I1206 03:34:00.954674 4980 flags.go:64] FLAG: --boot-id-file="/proc/sys/kernel/random/boot_id" Dec 06 03:34:00 crc kubenswrapper[4980]: I1206 03:34:00.954683 4980 flags.go:64] FLAG: --bootstrap-kubeconfig="/etc/kubernetes/kubeconfig" Dec 06 03:34:00 crc kubenswrapper[4980]: I1206 03:34:00.954691 4980 flags.go:64] FLAG: --cert-dir="/var/lib/kubelet/pki" Dec 06 03:34:00 crc kubenswrapper[4980]: I1206 03:34:00.954699 4980 flags.go:64] FLAG: --cgroup-driver="cgroupfs" Dec 06 03:34:00 crc kubenswrapper[4980]: I1206 03:34:00.954707 4980 flags.go:64] FLAG: --cgroup-root="" Dec 06 03:34:00 crc kubenswrapper[4980]: I1206 03:34:00.954715 4980 flags.go:64] FLAG: --cgroups-per-qos="true" Dec 06 03:34:00 crc kubenswrapper[4980]: I1206 03:34:00.954724 4980 flags.go:64] FLAG: --client-ca-file="" Dec 06 03:34:00 crc kubenswrapper[4980]: I1206 03:34:00.954731 4980 flags.go:64] FLAG: --cloud-config="" Dec 06 03:34:00 crc kubenswrapper[4980]: I1206 03:34:00.954741 4980 flags.go:64] FLAG: --cloud-provider="" Dec 06 03:34:00 crc kubenswrapper[4980]: I1206 03:34:00.954749 4980 flags.go:64] FLAG: --cluster-dns="[]" Dec 06 03:34:00 crc kubenswrapper[4980]: I1206 03:34:00.954766 4980 flags.go:64] FLAG: --cluster-domain="" Dec 06 03:34:00 crc kubenswrapper[4980]: I1206 03:34:00.954773 4980 flags.go:64] FLAG: --config="/etc/kubernetes/kubelet.conf" Dec 06 03:34:00 crc kubenswrapper[4980]: I1206 03:34:00.954781 4980 flags.go:64] FLAG: --config-dir="" Dec 06 03:34:00 crc kubenswrapper[4980]: I1206 03:34:00.954788 4980 flags.go:64] FLAG: --container-hints="/etc/cadvisor/container_hints.json" Dec 06 03:34:00 crc kubenswrapper[4980]: I1206 03:34:00.954796 4980 flags.go:64] FLAG: --container-log-max-files="5" Dec 06 03:34:00 crc kubenswrapper[4980]: I1206 03:34:00.954806 4980 flags.go:64] FLAG: --container-log-max-size="10Mi" Dec 06 03:34:00 crc kubenswrapper[4980]: I1206 03:34:00.954813 4980 flags.go:64] FLAG: --container-runtime-endpoint="/var/run/crio/crio.sock" Dec 06 03:34:00 crc kubenswrapper[4980]: I1206 03:34:00.954821 4980 flags.go:64] FLAG: --containerd="/run/containerd/containerd.sock" Dec 06 03:34:00 crc kubenswrapper[4980]: I1206 03:34:00.954829 4980 flags.go:64] FLAG: --containerd-namespace="k8s.io" Dec 06 03:34:00 crc kubenswrapper[4980]: I1206 03:34:00.954836 4980 flags.go:64] FLAG: --contention-profiling="false" Dec 06 03:34:00 crc kubenswrapper[4980]: I1206 03:34:00.954843 4980 flags.go:64] FLAG: --cpu-cfs-quota="true" Dec 06 03:34:00 crc kubenswrapper[4980]: I1206 03:34:00.954851 4980 flags.go:64] FLAG: --cpu-cfs-quota-period="100ms" Dec 06 03:34:00 crc kubenswrapper[4980]: I1206 03:34:00.954859 4980 flags.go:64] FLAG: --cpu-manager-policy="none" Dec 06 03:34:00 crc kubenswrapper[4980]: I1206 03:34:00.954865 4980 flags.go:64] FLAG: --cpu-manager-policy-options="" Dec 06 03:34:00 crc kubenswrapper[4980]: I1206 03:34:00.954875 4980 flags.go:64] FLAG: --cpu-manager-reconcile-period="10s" Dec 06 03:34:00 crc kubenswrapper[4980]: I1206 03:34:00.954882 4980 flags.go:64] FLAG: --enable-controller-attach-detach="true" Dec 06 03:34:00 crc kubenswrapper[4980]: I1206 03:34:00.954889 4980 flags.go:64] FLAG: --enable-debugging-handlers="true" Dec 06 03:34:00 crc kubenswrapper[4980]: I1206 03:34:00.954896 4980 flags.go:64] FLAG: --enable-load-reader="false" Dec 06 03:34:00 crc kubenswrapper[4980]: I1206 03:34:00.954905 4980 flags.go:64] FLAG: --enable-server="true" Dec 06 03:34:00 crc kubenswrapper[4980]: I1206 03:34:00.954913 4980 flags.go:64] FLAG: --enforce-node-allocatable="[pods]" Dec 06 03:34:00 crc kubenswrapper[4980]: I1206 03:34:00.954922 4980 flags.go:64] FLAG: --event-burst="100" Dec 06 03:34:00 crc kubenswrapper[4980]: I1206 03:34:00.954929 4980 flags.go:64] FLAG: --event-qps="50" Dec 06 03:34:00 crc kubenswrapper[4980]: I1206 03:34:00.954936 4980 flags.go:64] FLAG: --event-storage-age-limit="default=0" Dec 06 03:34:00 crc kubenswrapper[4980]: I1206 03:34:00.954943 4980 flags.go:64] FLAG: --event-storage-event-limit="default=0" Dec 06 03:34:00 crc kubenswrapper[4980]: I1206 03:34:00.954952 4980 flags.go:64] FLAG: --eviction-hard="" Dec 06 03:34:00 crc kubenswrapper[4980]: I1206 03:34:00.954962 4980 flags.go:64] FLAG: --eviction-max-pod-grace-period="0" Dec 06 03:34:00 crc kubenswrapper[4980]: I1206 03:34:00.954970 4980 flags.go:64] FLAG: --eviction-minimum-reclaim="" Dec 06 03:34:00 crc kubenswrapper[4980]: I1206 03:34:00.954977 4980 flags.go:64] FLAG: --eviction-pressure-transition-period="5m0s" Dec 06 03:34:00 crc kubenswrapper[4980]: I1206 03:34:00.954984 4980 flags.go:64] FLAG: --eviction-soft="" Dec 06 03:34:00 crc kubenswrapper[4980]: I1206 03:34:00.954992 4980 flags.go:64] FLAG: --eviction-soft-grace-period="" Dec 06 03:34:00 crc kubenswrapper[4980]: I1206 03:34:00.954999 4980 flags.go:64] FLAG: --exit-on-lock-contention="false" Dec 06 03:34:00 crc kubenswrapper[4980]: I1206 03:34:00.955007 4980 flags.go:64] FLAG: --experimental-allocatable-ignore-eviction="false" Dec 06 03:34:00 crc kubenswrapper[4980]: I1206 03:34:00.955016 4980 flags.go:64] FLAG: --experimental-mounter-path="" Dec 06 03:34:00 crc kubenswrapper[4980]: I1206 03:34:00.955024 4980 flags.go:64] FLAG: --fail-cgroupv1="false" Dec 06 03:34:00 crc kubenswrapper[4980]: I1206 03:34:00.955031 4980 flags.go:64] FLAG: --fail-swap-on="true" Dec 06 03:34:00 crc kubenswrapper[4980]: I1206 03:34:00.955038 4980 flags.go:64] FLAG: --feature-gates="" Dec 06 03:34:00 crc kubenswrapper[4980]: I1206 03:34:00.955048 4980 flags.go:64] FLAG: --file-check-frequency="20s" Dec 06 03:34:00 crc kubenswrapper[4980]: I1206 03:34:00.955056 4980 flags.go:64] FLAG: --global-housekeeping-interval="1m0s" Dec 06 03:34:00 crc kubenswrapper[4980]: I1206 03:34:00.955063 4980 flags.go:64] FLAG: --hairpin-mode="promiscuous-bridge" Dec 06 03:34:00 crc kubenswrapper[4980]: I1206 03:34:00.955071 4980 flags.go:64] FLAG: --healthz-bind-address="127.0.0.1" Dec 06 03:34:00 crc kubenswrapper[4980]: I1206 03:34:00.955079 4980 flags.go:64] FLAG: --healthz-port="10248" Dec 06 03:34:00 crc kubenswrapper[4980]: I1206 03:34:00.955086 4980 flags.go:64] FLAG: --help="false" Dec 06 03:34:00 crc kubenswrapper[4980]: I1206 03:34:00.955093 4980 flags.go:64] FLAG: --hostname-override="" Dec 06 03:34:00 crc kubenswrapper[4980]: I1206 03:34:00.955101 4980 flags.go:64] FLAG: --housekeeping-interval="10s" Dec 06 03:34:00 crc kubenswrapper[4980]: I1206 03:34:00.955109 4980 flags.go:64] FLAG: --http-check-frequency="20s" Dec 06 03:34:00 crc kubenswrapper[4980]: I1206 03:34:00.955117 4980 flags.go:64] FLAG: --image-credential-provider-bin-dir="" Dec 06 03:34:00 crc kubenswrapper[4980]: I1206 03:34:00.955124 4980 flags.go:64] FLAG: --image-credential-provider-config="" Dec 06 03:34:00 crc kubenswrapper[4980]: I1206 03:34:00.955131 4980 flags.go:64] FLAG: --image-gc-high-threshold="85" Dec 06 03:34:00 crc kubenswrapper[4980]: I1206 03:34:00.955138 4980 flags.go:64] FLAG: --image-gc-low-threshold="80" Dec 06 03:34:00 crc kubenswrapper[4980]: I1206 03:34:00.955145 4980 flags.go:64] FLAG: --image-service-endpoint="" Dec 06 03:34:00 crc kubenswrapper[4980]: I1206 03:34:00.955152 4980 flags.go:64] FLAG: --kernel-memcg-notification="false" Dec 06 03:34:00 crc kubenswrapper[4980]: I1206 03:34:00.955159 4980 flags.go:64] FLAG: --kube-api-burst="100" Dec 06 03:34:00 crc kubenswrapper[4980]: I1206 03:34:00.955167 4980 flags.go:64] FLAG: --kube-api-content-type="application/vnd.kubernetes.protobuf" Dec 06 03:34:00 crc kubenswrapper[4980]: I1206 03:34:00.955174 4980 flags.go:64] FLAG: --kube-api-qps="50" Dec 06 03:34:00 crc kubenswrapper[4980]: I1206 03:34:00.955183 4980 flags.go:64] FLAG: --kube-reserved="" Dec 06 03:34:00 crc kubenswrapper[4980]: I1206 03:34:00.955191 4980 flags.go:64] FLAG: --kube-reserved-cgroup="" Dec 06 03:34:00 crc kubenswrapper[4980]: I1206 03:34:00.955198 4980 flags.go:64] FLAG: --kubeconfig="/var/lib/kubelet/kubeconfig" Dec 06 03:34:00 crc kubenswrapper[4980]: I1206 03:34:00.955206 4980 flags.go:64] FLAG: --kubelet-cgroups="" Dec 06 03:34:00 crc kubenswrapper[4980]: I1206 03:34:00.955213 4980 flags.go:64] FLAG: --local-storage-capacity-isolation="true" Dec 06 03:34:00 crc kubenswrapper[4980]: I1206 03:34:00.955220 4980 flags.go:64] FLAG: --lock-file="" Dec 06 03:34:00 crc kubenswrapper[4980]: I1206 03:34:00.955227 4980 flags.go:64] FLAG: --log-cadvisor-usage="false" Dec 06 03:34:00 crc kubenswrapper[4980]: I1206 03:34:00.955236 4980 flags.go:64] FLAG: --log-flush-frequency="5s" Dec 06 03:34:00 crc kubenswrapper[4980]: I1206 03:34:00.955243 4980 flags.go:64] FLAG: --log-json-info-buffer-size="0" Dec 06 03:34:00 crc kubenswrapper[4980]: I1206 03:34:00.955254 4980 flags.go:64] FLAG: --log-json-split-stream="false" Dec 06 03:34:00 crc kubenswrapper[4980]: I1206 03:34:00.955262 4980 flags.go:64] FLAG: --log-text-info-buffer-size="0" Dec 06 03:34:00 crc kubenswrapper[4980]: I1206 03:34:00.955269 4980 flags.go:64] FLAG: --log-text-split-stream="false" Dec 06 03:34:00 crc kubenswrapper[4980]: I1206 03:34:00.955277 4980 flags.go:64] FLAG: --logging-format="text" Dec 06 03:34:00 crc kubenswrapper[4980]: I1206 03:34:00.955284 4980 flags.go:64] FLAG: --machine-id-file="/etc/machine-id,/var/lib/dbus/machine-id" Dec 06 03:34:00 crc kubenswrapper[4980]: I1206 03:34:00.955292 4980 flags.go:64] FLAG: --make-iptables-util-chains="true" Dec 06 03:34:00 crc kubenswrapper[4980]: I1206 03:34:00.955299 4980 flags.go:64] FLAG: --manifest-url="" Dec 06 03:34:00 crc kubenswrapper[4980]: I1206 03:34:00.955307 4980 flags.go:64] FLAG: --manifest-url-header="" Dec 06 03:34:00 crc kubenswrapper[4980]: I1206 03:34:00.955316 4980 flags.go:64] FLAG: --max-housekeeping-interval="15s" Dec 06 03:34:00 crc kubenswrapper[4980]: I1206 03:34:00.955324 4980 flags.go:64] FLAG: --max-open-files="1000000" Dec 06 03:34:00 crc kubenswrapper[4980]: I1206 03:34:00.955332 4980 flags.go:64] FLAG: --max-pods="110" Dec 06 03:34:00 crc kubenswrapper[4980]: I1206 03:34:00.955340 4980 flags.go:64] FLAG: --maximum-dead-containers="-1" Dec 06 03:34:00 crc kubenswrapper[4980]: I1206 03:34:00.955348 4980 flags.go:64] FLAG: --maximum-dead-containers-per-container="1" Dec 06 03:34:00 crc kubenswrapper[4980]: I1206 03:34:00.955356 4980 flags.go:64] FLAG: --memory-manager-policy="None" Dec 06 03:34:00 crc kubenswrapper[4980]: I1206 03:34:00.955363 4980 flags.go:64] FLAG: --minimum-container-ttl-duration="6m0s" Dec 06 03:34:00 crc kubenswrapper[4980]: I1206 03:34:00.955370 4980 flags.go:64] FLAG: --minimum-image-ttl-duration="2m0s" Dec 06 03:34:00 crc kubenswrapper[4980]: I1206 03:34:00.955377 4980 flags.go:64] FLAG: --node-ip="192.168.126.11" Dec 06 03:34:00 crc kubenswrapper[4980]: I1206 03:34:00.955385 4980 flags.go:64] FLAG: --node-labels="node-role.kubernetes.io/control-plane=,node-role.kubernetes.io/master=,node.openshift.io/os_id=rhcos" Dec 06 03:34:00 crc kubenswrapper[4980]: I1206 03:34:00.955408 4980 flags.go:64] FLAG: --node-status-max-images="50" Dec 06 03:34:00 crc kubenswrapper[4980]: I1206 03:34:00.955416 4980 flags.go:64] FLAG: --node-status-update-frequency="10s" Dec 06 03:34:00 crc kubenswrapper[4980]: I1206 03:34:00.955423 4980 flags.go:64] FLAG: --oom-score-adj="-999" Dec 06 03:34:00 crc kubenswrapper[4980]: I1206 03:34:00.955430 4980 flags.go:64] FLAG: --pod-cidr="" Dec 06 03:34:00 crc kubenswrapper[4980]: I1206 03:34:00.955437 4980 flags.go:64] FLAG: --pod-infra-container-image="quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:33549946e22a9ffa738fd94b1345f90921bc8f92fa6137784cb33c77ad806f9d" Dec 06 03:34:00 crc kubenswrapper[4980]: I1206 03:34:00.955452 4980 flags.go:64] FLAG: --pod-manifest-path="" Dec 06 03:34:00 crc kubenswrapper[4980]: I1206 03:34:00.955460 4980 flags.go:64] FLAG: --pod-max-pids="-1" Dec 06 03:34:00 crc kubenswrapper[4980]: I1206 03:34:00.955469 4980 flags.go:64] FLAG: --pods-per-core="0" Dec 06 03:34:00 crc kubenswrapper[4980]: I1206 03:34:00.955477 4980 flags.go:64] FLAG: --port="10250" Dec 06 03:34:00 crc kubenswrapper[4980]: I1206 03:34:00.955486 4980 flags.go:64] FLAG: --protect-kernel-defaults="false" Dec 06 03:34:00 crc kubenswrapper[4980]: I1206 03:34:00.955494 4980 flags.go:64] FLAG: --provider-id="" Dec 06 03:34:00 crc kubenswrapper[4980]: I1206 03:34:00.955501 4980 flags.go:64] FLAG: --qos-reserved="" Dec 06 03:34:00 crc kubenswrapper[4980]: I1206 03:34:00.955508 4980 flags.go:64] FLAG: --read-only-port="10255" Dec 06 03:34:00 crc kubenswrapper[4980]: I1206 03:34:00.955545 4980 flags.go:64] FLAG: --register-node="true" Dec 06 03:34:00 crc kubenswrapper[4980]: I1206 03:34:00.955553 4980 flags.go:64] FLAG: --register-schedulable="true" Dec 06 03:34:00 crc kubenswrapper[4980]: I1206 03:34:00.955561 4980 flags.go:64] FLAG: --register-with-taints="node-role.kubernetes.io/master=:NoSchedule" Dec 06 03:34:00 crc kubenswrapper[4980]: I1206 03:34:00.955574 4980 flags.go:64] FLAG: --registry-burst="10" Dec 06 03:34:00 crc kubenswrapper[4980]: I1206 03:34:00.955581 4980 flags.go:64] FLAG: --registry-qps="5" Dec 06 03:34:00 crc kubenswrapper[4980]: I1206 03:34:00.955588 4980 flags.go:64] FLAG: --reserved-cpus="" Dec 06 03:34:00 crc kubenswrapper[4980]: I1206 03:34:00.955595 4980 flags.go:64] FLAG: --reserved-memory="" Dec 06 03:34:00 crc kubenswrapper[4980]: I1206 03:34:00.955604 4980 flags.go:64] FLAG: --resolv-conf="/etc/resolv.conf" Dec 06 03:34:00 crc kubenswrapper[4980]: I1206 03:34:00.955612 4980 flags.go:64] FLAG: --root-dir="/var/lib/kubelet" Dec 06 03:34:00 crc kubenswrapper[4980]: I1206 03:34:00.955619 4980 flags.go:64] FLAG: --rotate-certificates="false" Dec 06 03:34:00 crc kubenswrapper[4980]: I1206 03:34:00.955627 4980 flags.go:64] FLAG: --rotate-server-certificates="false" Dec 06 03:34:00 crc kubenswrapper[4980]: I1206 03:34:00.955634 4980 flags.go:64] FLAG: --runonce="false" Dec 06 03:34:00 crc kubenswrapper[4980]: I1206 03:34:00.955641 4980 flags.go:64] FLAG: --runtime-cgroups="/system.slice/crio.service" Dec 06 03:34:00 crc kubenswrapper[4980]: I1206 03:34:00.955649 4980 flags.go:64] FLAG: --runtime-request-timeout="2m0s" Dec 06 03:34:00 crc kubenswrapper[4980]: I1206 03:34:00.955658 4980 flags.go:64] FLAG: --seccomp-default="false" Dec 06 03:34:00 crc kubenswrapper[4980]: I1206 03:34:00.955665 4980 flags.go:64] FLAG: --serialize-image-pulls="true" Dec 06 03:34:00 crc kubenswrapper[4980]: I1206 03:34:00.955672 4980 flags.go:64] FLAG: --storage-driver-buffer-duration="1m0s" Dec 06 03:34:00 crc kubenswrapper[4980]: I1206 03:34:00.955679 4980 flags.go:64] FLAG: --storage-driver-db="cadvisor" Dec 06 03:34:00 crc kubenswrapper[4980]: I1206 03:34:00.955687 4980 flags.go:64] FLAG: --storage-driver-host="localhost:8086" Dec 06 03:34:00 crc kubenswrapper[4980]: I1206 03:34:00.955695 4980 flags.go:64] FLAG: --storage-driver-password="root" Dec 06 03:34:00 crc kubenswrapper[4980]: I1206 03:34:00.955701 4980 flags.go:64] FLAG: --storage-driver-secure="false" Dec 06 03:34:00 crc kubenswrapper[4980]: I1206 03:34:00.955708 4980 flags.go:64] FLAG: --storage-driver-table="stats" Dec 06 03:34:00 crc kubenswrapper[4980]: I1206 03:34:00.955717 4980 flags.go:64] FLAG: --storage-driver-user="root" Dec 06 03:34:00 crc kubenswrapper[4980]: I1206 03:34:00.955724 4980 flags.go:64] FLAG: --streaming-connection-idle-timeout="4h0m0s" Dec 06 03:34:00 crc kubenswrapper[4980]: I1206 03:34:00.955732 4980 flags.go:64] FLAG: --sync-frequency="1m0s" Dec 06 03:34:00 crc kubenswrapper[4980]: I1206 03:34:00.955741 4980 flags.go:64] FLAG: --system-cgroups="" Dec 06 03:34:00 crc kubenswrapper[4980]: I1206 03:34:00.955748 4980 flags.go:64] FLAG: --system-reserved="cpu=200m,ephemeral-storage=350Mi,memory=350Mi" Dec 06 03:34:00 crc kubenswrapper[4980]: I1206 03:34:00.955759 4980 flags.go:64] FLAG: --system-reserved-cgroup="" Dec 06 03:34:00 crc kubenswrapper[4980]: I1206 03:34:00.955767 4980 flags.go:64] FLAG: --tls-cert-file="" Dec 06 03:34:00 crc kubenswrapper[4980]: I1206 03:34:00.955775 4980 flags.go:64] FLAG: --tls-cipher-suites="[]" Dec 06 03:34:00 crc kubenswrapper[4980]: I1206 03:34:00.955786 4980 flags.go:64] FLAG: --tls-min-version="" Dec 06 03:34:00 crc kubenswrapper[4980]: I1206 03:34:00.955793 4980 flags.go:64] FLAG: --tls-private-key-file="" Dec 06 03:34:00 crc kubenswrapper[4980]: I1206 03:34:00.955801 4980 flags.go:64] FLAG: --topology-manager-policy="none" Dec 06 03:34:00 crc kubenswrapper[4980]: I1206 03:34:00.955809 4980 flags.go:64] FLAG: --topology-manager-policy-options="" Dec 06 03:34:00 crc kubenswrapper[4980]: I1206 03:34:00.955817 4980 flags.go:64] FLAG: --topology-manager-scope="container" Dec 06 03:34:00 crc kubenswrapper[4980]: I1206 03:34:00.955825 4980 flags.go:64] FLAG: --v="2" Dec 06 03:34:00 crc kubenswrapper[4980]: I1206 03:34:00.955835 4980 flags.go:64] FLAG: --version="false" Dec 06 03:34:00 crc kubenswrapper[4980]: I1206 03:34:00.955843 4980 flags.go:64] FLAG: --vmodule="" Dec 06 03:34:00 crc kubenswrapper[4980]: I1206 03:34:00.955852 4980 flags.go:64] FLAG: --volume-plugin-dir="/etc/kubernetes/kubelet-plugins/volume/exec" Dec 06 03:34:00 crc kubenswrapper[4980]: I1206 03:34:00.955860 4980 flags.go:64] FLAG: --volume-stats-agg-period="1m0s" Dec 06 03:34:00 crc kubenswrapper[4980]: W1206 03:34:00.956032 4980 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Dec 06 03:34:00 crc kubenswrapper[4980]: W1206 03:34:00.956042 4980 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Dec 06 03:34:00 crc kubenswrapper[4980]: W1206 03:34:00.956049 4980 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Dec 06 03:34:00 crc kubenswrapper[4980]: W1206 03:34:00.956056 4980 feature_gate.go:330] unrecognized feature gate: PlatformOperators Dec 06 03:34:00 crc kubenswrapper[4980]: W1206 03:34:00.956062 4980 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Dec 06 03:34:00 crc kubenswrapper[4980]: W1206 03:34:00.956069 4980 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Dec 06 03:34:00 crc kubenswrapper[4980]: W1206 03:34:00.956083 4980 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Dec 06 03:34:00 crc kubenswrapper[4980]: W1206 03:34:00.956090 4980 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Dec 06 03:34:00 crc kubenswrapper[4980]: W1206 03:34:00.956095 4980 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Dec 06 03:34:00 crc kubenswrapper[4980]: W1206 03:34:00.956101 4980 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Dec 06 03:34:00 crc kubenswrapper[4980]: W1206 03:34:00.956108 4980 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Dec 06 03:34:00 crc kubenswrapper[4980]: W1206 03:34:00.956114 4980 feature_gate.go:330] unrecognized feature gate: PinnedImages Dec 06 03:34:00 crc kubenswrapper[4980]: W1206 03:34:00.956121 4980 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Dec 06 03:34:00 crc kubenswrapper[4980]: W1206 03:34:00.956127 4980 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Dec 06 03:34:00 crc kubenswrapper[4980]: W1206 03:34:00.956133 4980 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Dec 06 03:34:00 crc kubenswrapper[4980]: W1206 03:34:00.956140 4980 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Dec 06 03:34:00 crc kubenswrapper[4980]: W1206 03:34:00.956146 4980 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Dec 06 03:34:00 crc kubenswrapper[4980]: W1206 03:34:00.956154 4980 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Dec 06 03:34:00 crc kubenswrapper[4980]: W1206 03:34:00.956164 4980 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Dec 06 03:34:00 crc kubenswrapper[4980]: W1206 03:34:00.956172 4980 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Dec 06 03:34:00 crc kubenswrapper[4980]: W1206 03:34:00.956180 4980 feature_gate.go:330] unrecognized feature gate: InsightsConfig Dec 06 03:34:00 crc kubenswrapper[4980]: W1206 03:34:00.956187 4980 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Dec 06 03:34:00 crc kubenswrapper[4980]: W1206 03:34:00.956194 4980 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Dec 06 03:34:00 crc kubenswrapper[4980]: W1206 03:34:00.956201 4980 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Dec 06 03:34:00 crc kubenswrapper[4980]: W1206 03:34:00.956207 4980 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Dec 06 03:34:00 crc kubenswrapper[4980]: W1206 03:34:00.956215 4980 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Dec 06 03:34:00 crc kubenswrapper[4980]: W1206 03:34:00.956221 4980 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Dec 06 03:34:00 crc kubenswrapper[4980]: W1206 03:34:00.956227 4980 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Dec 06 03:34:00 crc kubenswrapper[4980]: W1206 03:34:00.956235 4980 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Dec 06 03:34:00 crc kubenswrapper[4980]: W1206 03:34:00.956242 4980 feature_gate.go:330] unrecognized feature gate: OVNObservability Dec 06 03:34:00 crc kubenswrapper[4980]: W1206 03:34:00.956248 4980 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Dec 06 03:34:00 crc kubenswrapper[4980]: W1206 03:34:00.956255 4980 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Dec 06 03:34:00 crc kubenswrapper[4980]: W1206 03:34:00.956261 4980 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Dec 06 03:34:00 crc kubenswrapper[4980]: W1206 03:34:00.956267 4980 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Dec 06 03:34:00 crc kubenswrapper[4980]: W1206 03:34:00.956273 4980 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Dec 06 03:34:00 crc kubenswrapper[4980]: W1206 03:34:00.956278 4980 feature_gate.go:330] unrecognized feature gate: GatewayAPI Dec 06 03:34:00 crc kubenswrapper[4980]: W1206 03:34:00.956284 4980 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Dec 06 03:34:00 crc kubenswrapper[4980]: W1206 03:34:00.956290 4980 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Dec 06 03:34:00 crc kubenswrapper[4980]: W1206 03:34:00.956299 4980 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Dec 06 03:34:00 crc kubenswrapper[4980]: W1206 03:34:00.956306 4980 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Dec 06 03:34:00 crc kubenswrapper[4980]: W1206 03:34:00.956312 4980 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Dec 06 03:34:00 crc kubenswrapper[4980]: W1206 03:34:00.956319 4980 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Dec 06 03:34:00 crc kubenswrapper[4980]: W1206 03:34:00.956326 4980 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Dec 06 03:34:00 crc kubenswrapper[4980]: W1206 03:34:00.956332 4980 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Dec 06 03:34:00 crc kubenswrapper[4980]: W1206 03:34:00.956338 4980 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Dec 06 03:34:00 crc kubenswrapper[4980]: W1206 03:34:00.956344 4980 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Dec 06 03:34:00 crc kubenswrapper[4980]: W1206 03:34:00.956350 4980 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Dec 06 03:34:00 crc kubenswrapper[4980]: W1206 03:34:00.956356 4980 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Dec 06 03:34:00 crc kubenswrapper[4980]: W1206 03:34:00.956362 4980 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Dec 06 03:34:00 crc kubenswrapper[4980]: W1206 03:34:00.956368 4980 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Dec 06 03:34:00 crc kubenswrapper[4980]: W1206 03:34:00.956374 4980 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Dec 06 03:34:00 crc kubenswrapper[4980]: W1206 03:34:00.956381 4980 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Dec 06 03:34:00 crc kubenswrapper[4980]: W1206 03:34:00.956387 4980 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Dec 06 03:34:00 crc kubenswrapper[4980]: W1206 03:34:00.956394 4980 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Dec 06 03:34:00 crc kubenswrapper[4980]: W1206 03:34:00.956400 4980 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Dec 06 03:34:00 crc kubenswrapper[4980]: W1206 03:34:00.956406 4980 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Dec 06 03:34:00 crc kubenswrapper[4980]: W1206 03:34:00.956412 4980 feature_gate.go:330] unrecognized feature gate: Example Dec 06 03:34:00 crc kubenswrapper[4980]: W1206 03:34:00.956418 4980 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Dec 06 03:34:00 crc kubenswrapper[4980]: W1206 03:34:00.956426 4980 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Dec 06 03:34:00 crc kubenswrapper[4980]: W1206 03:34:00.956434 4980 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Dec 06 03:34:00 crc kubenswrapper[4980]: W1206 03:34:00.956442 4980 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Dec 06 03:34:00 crc kubenswrapper[4980]: W1206 03:34:00.956450 4980 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Dec 06 03:34:00 crc kubenswrapper[4980]: W1206 03:34:00.956458 4980 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Dec 06 03:34:00 crc kubenswrapper[4980]: W1206 03:34:00.956465 4980 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Dec 06 03:34:00 crc kubenswrapper[4980]: W1206 03:34:00.956473 4980 feature_gate.go:330] unrecognized feature gate: NewOLM Dec 06 03:34:00 crc kubenswrapper[4980]: W1206 03:34:00.956480 4980 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Dec 06 03:34:00 crc kubenswrapper[4980]: W1206 03:34:00.956486 4980 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Dec 06 03:34:00 crc kubenswrapper[4980]: W1206 03:34:00.956493 4980 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Dec 06 03:34:00 crc kubenswrapper[4980]: W1206 03:34:00.956500 4980 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Dec 06 03:34:00 crc kubenswrapper[4980]: W1206 03:34:00.956506 4980 feature_gate.go:330] unrecognized feature gate: SignatureStores Dec 06 03:34:00 crc kubenswrapper[4980]: W1206 03:34:00.956541 4980 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Dec 06 03:34:00 crc kubenswrapper[4980]: I1206 03:34:00.956560 4980 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Dec 06 03:34:00 crc kubenswrapper[4980]: I1206 03:34:00.979708 4980 server.go:491] "Kubelet version" kubeletVersion="v1.31.5" Dec 06 03:34:00 crc kubenswrapper[4980]: I1206 03:34:00.979791 4980 server.go:493] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Dec 06 03:34:00 crc kubenswrapper[4980]: W1206 03:34:00.979921 4980 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Dec 06 03:34:00 crc kubenswrapper[4980]: W1206 03:34:00.979943 4980 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Dec 06 03:34:00 crc kubenswrapper[4980]: W1206 03:34:00.979950 4980 feature_gate.go:330] unrecognized feature gate: InsightsConfig Dec 06 03:34:00 crc kubenswrapper[4980]: W1206 03:34:00.979958 4980 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Dec 06 03:34:00 crc kubenswrapper[4980]: W1206 03:34:00.979968 4980 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Dec 06 03:34:00 crc kubenswrapper[4980]: W1206 03:34:00.979980 4980 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Dec 06 03:34:00 crc kubenswrapper[4980]: W1206 03:34:00.979987 4980 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Dec 06 03:34:00 crc kubenswrapper[4980]: W1206 03:34:00.979993 4980 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Dec 06 03:34:00 crc kubenswrapper[4980]: W1206 03:34:00.979999 4980 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Dec 06 03:34:00 crc kubenswrapper[4980]: W1206 03:34:00.980004 4980 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Dec 06 03:34:00 crc kubenswrapper[4980]: W1206 03:34:00.980010 4980 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Dec 06 03:34:00 crc kubenswrapper[4980]: W1206 03:34:00.980016 4980 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Dec 06 03:34:00 crc kubenswrapper[4980]: W1206 03:34:00.980021 4980 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Dec 06 03:34:00 crc kubenswrapper[4980]: W1206 03:34:00.980027 4980 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Dec 06 03:34:00 crc kubenswrapper[4980]: W1206 03:34:00.980033 4980 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Dec 06 03:34:00 crc kubenswrapper[4980]: W1206 03:34:00.980039 4980 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Dec 06 03:34:00 crc kubenswrapper[4980]: W1206 03:34:00.980045 4980 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Dec 06 03:34:00 crc kubenswrapper[4980]: W1206 03:34:00.980051 4980 feature_gate.go:330] unrecognized feature gate: OVNObservability Dec 06 03:34:00 crc kubenswrapper[4980]: W1206 03:34:00.980056 4980 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Dec 06 03:34:00 crc kubenswrapper[4980]: W1206 03:34:00.980062 4980 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Dec 06 03:34:00 crc kubenswrapper[4980]: W1206 03:34:00.980068 4980 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Dec 06 03:34:00 crc kubenswrapper[4980]: W1206 03:34:00.980074 4980 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Dec 06 03:34:00 crc kubenswrapper[4980]: W1206 03:34:00.980079 4980 feature_gate.go:330] unrecognized feature gate: PlatformOperators Dec 06 03:34:00 crc kubenswrapper[4980]: W1206 03:34:00.980085 4980 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Dec 06 03:34:00 crc kubenswrapper[4980]: W1206 03:34:00.980090 4980 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Dec 06 03:34:00 crc kubenswrapper[4980]: W1206 03:34:00.980095 4980 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Dec 06 03:34:00 crc kubenswrapper[4980]: W1206 03:34:00.980102 4980 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Dec 06 03:34:00 crc kubenswrapper[4980]: W1206 03:34:00.980107 4980 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Dec 06 03:34:00 crc kubenswrapper[4980]: W1206 03:34:00.980112 4980 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Dec 06 03:34:00 crc kubenswrapper[4980]: W1206 03:34:00.980118 4980 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Dec 06 03:34:00 crc kubenswrapper[4980]: W1206 03:34:00.980123 4980 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Dec 06 03:34:00 crc kubenswrapper[4980]: W1206 03:34:00.980128 4980 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Dec 06 03:34:00 crc kubenswrapper[4980]: W1206 03:34:00.980135 4980 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Dec 06 03:34:00 crc kubenswrapper[4980]: W1206 03:34:00.980143 4980 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Dec 06 03:34:00 crc kubenswrapper[4980]: W1206 03:34:00.980152 4980 feature_gate.go:330] unrecognized feature gate: Example Dec 06 03:34:00 crc kubenswrapper[4980]: W1206 03:34:00.980158 4980 feature_gate.go:330] unrecognized feature gate: SignatureStores Dec 06 03:34:00 crc kubenswrapper[4980]: W1206 03:34:00.980165 4980 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Dec 06 03:34:00 crc kubenswrapper[4980]: W1206 03:34:00.980171 4980 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Dec 06 03:34:00 crc kubenswrapper[4980]: W1206 03:34:00.980177 4980 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Dec 06 03:34:00 crc kubenswrapper[4980]: W1206 03:34:00.980183 4980 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Dec 06 03:34:00 crc kubenswrapper[4980]: W1206 03:34:00.980189 4980 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Dec 06 03:34:00 crc kubenswrapper[4980]: W1206 03:34:00.980194 4980 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Dec 06 03:34:00 crc kubenswrapper[4980]: W1206 03:34:00.980200 4980 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Dec 06 03:34:00 crc kubenswrapper[4980]: W1206 03:34:00.980205 4980 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Dec 06 03:34:00 crc kubenswrapper[4980]: W1206 03:34:00.980211 4980 feature_gate.go:330] unrecognized feature gate: GatewayAPI Dec 06 03:34:00 crc kubenswrapper[4980]: W1206 03:34:00.980216 4980 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Dec 06 03:34:00 crc kubenswrapper[4980]: W1206 03:34:00.980222 4980 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Dec 06 03:34:00 crc kubenswrapper[4980]: W1206 03:34:00.980227 4980 feature_gate.go:330] unrecognized feature gate: NewOLM Dec 06 03:34:00 crc kubenswrapper[4980]: W1206 03:34:00.980234 4980 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Dec 06 03:34:00 crc kubenswrapper[4980]: W1206 03:34:00.980240 4980 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Dec 06 03:34:00 crc kubenswrapper[4980]: W1206 03:34:00.980247 4980 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Dec 06 03:34:00 crc kubenswrapper[4980]: W1206 03:34:00.980253 4980 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Dec 06 03:34:00 crc kubenswrapper[4980]: W1206 03:34:00.980259 4980 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Dec 06 03:34:00 crc kubenswrapper[4980]: W1206 03:34:00.980264 4980 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Dec 06 03:34:00 crc kubenswrapper[4980]: W1206 03:34:00.980270 4980 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Dec 06 03:34:00 crc kubenswrapper[4980]: W1206 03:34:00.980276 4980 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Dec 06 03:34:00 crc kubenswrapper[4980]: W1206 03:34:00.980281 4980 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Dec 06 03:34:00 crc kubenswrapper[4980]: W1206 03:34:00.980286 4980 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Dec 06 03:34:00 crc kubenswrapper[4980]: W1206 03:34:00.980291 4980 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Dec 06 03:34:00 crc kubenswrapper[4980]: W1206 03:34:00.980298 4980 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Dec 06 03:34:00 crc kubenswrapper[4980]: W1206 03:34:00.980305 4980 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Dec 06 03:34:00 crc kubenswrapper[4980]: W1206 03:34:00.980312 4980 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Dec 06 03:34:00 crc kubenswrapper[4980]: W1206 03:34:00.980318 4980 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Dec 06 03:34:00 crc kubenswrapper[4980]: W1206 03:34:00.980324 4980 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Dec 06 03:34:00 crc kubenswrapper[4980]: W1206 03:34:00.980330 4980 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Dec 06 03:34:00 crc kubenswrapper[4980]: W1206 03:34:00.980336 4980 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Dec 06 03:34:00 crc kubenswrapper[4980]: W1206 03:34:00.980341 4980 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Dec 06 03:34:00 crc kubenswrapper[4980]: W1206 03:34:00.980347 4980 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Dec 06 03:34:00 crc kubenswrapper[4980]: W1206 03:34:00.980352 4980 feature_gate.go:330] unrecognized feature gate: PinnedImages Dec 06 03:34:00 crc kubenswrapper[4980]: W1206 03:34:00.980358 4980 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Dec 06 03:34:00 crc kubenswrapper[4980]: W1206 03:34:00.980371 4980 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Dec 06 03:34:00 crc kubenswrapper[4980]: I1206 03:34:00.980381 4980 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Dec 06 03:34:00 crc kubenswrapper[4980]: W1206 03:34:00.980647 4980 feature_gate.go:330] unrecognized feature gate: SignatureStores Dec 06 03:34:00 crc kubenswrapper[4980]: W1206 03:34:00.980677 4980 feature_gate.go:330] unrecognized feature gate: PlatformOperators Dec 06 03:34:00 crc kubenswrapper[4980]: W1206 03:34:00.980684 4980 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Dec 06 03:34:00 crc kubenswrapper[4980]: W1206 03:34:00.980690 4980 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Dec 06 03:34:00 crc kubenswrapper[4980]: W1206 03:34:00.980695 4980 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Dec 06 03:34:00 crc kubenswrapper[4980]: W1206 03:34:00.980701 4980 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Dec 06 03:34:00 crc kubenswrapper[4980]: W1206 03:34:00.980706 4980 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Dec 06 03:34:00 crc kubenswrapper[4980]: W1206 03:34:00.980711 4980 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Dec 06 03:34:00 crc kubenswrapper[4980]: W1206 03:34:00.980716 4980 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Dec 06 03:34:00 crc kubenswrapper[4980]: W1206 03:34:00.980722 4980 feature_gate.go:330] unrecognized feature gate: Example Dec 06 03:34:00 crc kubenswrapper[4980]: W1206 03:34:00.980727 4980 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Dec 06 03:34:00 crc kubenswrapper[4980]: W1206 03:34:00.980732 4980 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Dec 06 03:34:00 crc kubenswrapper[4980]: W1206 03:34:00.980737 4980 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Dec 06 03:34:00 crc kubenswrapper[4980]: W1206 03:34:00.980743 4980 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Dec 06 03:34:00 crc kubenswrapper[4980]: W1206 03:34:00.980748 4980 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Dec 06 03:34:00 crc kubenswrapper[4980]: W1206 03:34:00.980754 4980 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Dec 06 03:34:00 crc kubenswrapper[4980]: W1206 03:34:00.980759 4980 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Dec 06 03:34:00 crc kubenswrapper[4980]: W1206 03:34:00.980764 4980 feature_gate.go:330] unrecognized feature gate: OVNObservability Dec 06 03:34:00 crc kubenswrapper[4980]: W1206 03:34:00.980770 4980 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Dec 06 03:34:00 crc kubenswrapper[4980]: W1206 03:34:00.980775 4980 feature_gate.go:330] unrecognized feature gate: InsightsConfig Dec 06 03:34:00 crc kubenswrapper[4980]: W1206 03:34:00.980780 4980 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Dec 06 03:34:00 crc kubenswrapper[4980]: W1206 03:34:00.980788 4980 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Dec 06 03:34:00 crc kubenswrapper[4980]: W1206 03:34:00.980794 4980 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Dec 06 03:34:00 crc kubenswrapper[4980]: W1206 03:34:00.980801 4980 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Dec 06 03:34:00 crc kubenswrapper[4980]: W1206 03:34:00.980809 4980 feature_gate.go:330] unrecognized feature gate: GatewayAPI Dec 06 03:34:00 crc kubenswrapper[4980]: W1206 03:34:00.980816 4980 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Dec 06 03:34:00 crc kubenswrapper[4980]: W1206 03:34:00.980821 4980 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Dec 06 03:34:00 crc kubenswrapper[4980]: W1206 03:34:00.980828 4980 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Dec 06 03:34:00 crc kubenswrapper[4980]: W1206 03:34:00.980834 4980 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Dec 06 03:34:00 crc kubenswrapper[4980]: W1206 03:34:00.980840 4980 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Dec 06 03:34:00 crc kubenswrapper[4980]: W1206 03:34:00.980846 4980 feature_gate.go:330] unrecognized feature gate: PinnedImages Dec 06 03:34:00 crc kubenswrapper[4980]: W1206 03:34:00.980851 4980 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Dec 06 03:34:00 crc kubenswrapper[4980]: W1206 03:34:00.980857 4980 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Dec 06 03:34:00 crc kubenswrapper[4980]: W1206 03:34:00.980863 4980 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Dec 06 03:34:00 crc kubenswrapper[4980]: W1206 03:34:00.980896 4980 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Dec 06 03:34:00 crc kubenswrapper[4980]: W1206 03:34:00.980903 4980 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Dec 06 03:34:00 crc kubenswrapper[4980]: W1206 03:34:00.980909 4980 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Dec 06 03:34:00 crc kubenswrapper[4980]: W1206 03:34:00.980915 4980 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Dec 06 03:34:00 crc kubenswrapper[4980]: W1206 03:34:00.980921 4980 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Dec 06 03:34:00 crc kubenswrapper[4980]: W1206 03:34:00.980927 4980 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Dec 06 03:34:00 crc kubenswrapper[4980]: W1206 03:34:00.980934 4980 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Dec 06 03:34:00 crc kubenswrapper[4980]: W1206 03:34:00.980941 4980 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Dec 06 03:34:00 crc kubenswrapper[4980]: W1206 03:34:00.980947 4980 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Dec 06 03:34:00 crc kubenswrapper[4980]: W1206 03:34:00.980953 4980 feature_gate.go:330] unrecognized feature gate: NewOLM Dec 06 03:34:00 crc kubenswrapper[4980]: W1206 03:34:00.980959 4980 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Dec 06 03:34:00 crc kubenswrapper[4980]: W1206 03:34:00.980966 4980 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Dec 06 03:34:00 crc kubenswrapper[4980]: W1206 03:34:00.980972 4980 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Dec 06 03:34:00 crc kubenswrapper[4980]: W1206 03:34:00.980978 4980 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Dec 06 03:34:00 crc kubenswrapper[4980]: W1206 03:34:00.980984 4980 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Dec 06 03:34:00 crc kubenswrapper[4980]: W1206 03:34:00.980989 4980 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Dec 06 03:34:00 crc kubenswrapper[4980]: W1206 03:34:00.980994 4980 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Dec 06 03:34:00 crc kubenswrapper[4980]: W1206 03:34:00.981000 4980 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Dec 06 03:34:00 crc kubenswrapper[4980]: W1206 03:34:00.981005 4980 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Dec 06 03:34:00 crc kubenswrapper[4980]: W1206 03:34:00.981010 4980 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Dec 06 03:34:00 crc kubenswrapper[4980]: W1206 03:34:00.981015 4980 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Dec 06 03:34:00 crc kubenswrapper[4980]: W1206 03:34:00.981022 4980 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Dec 06 03:34:00 crc kubenswrapper[4980]: W1206 03:34:00.981029 4980 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Dec 06 03:34:00 crc kubenswrapper[4980]: W1206 03:34:00.981035 4980 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Dec 06 03:34:00 crc kubenswrapper[4980]: W1206 03:34:00.981040 4980 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Dec 06 03:34:00 crc kubenswrapper[4980]: W1206 03:34:00.981045 4980 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Dec 06 03:34:00 crc kubenswrapper[4980]: W1206 03:34:00.981050 4980 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Dec 06 03:34:00 crc kubenswrapper[4980]: W1206 03:34:00.981055 4980 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Dec 06 03:34:00 crc kubenswrapper[4980]: W1206 03:34:00.981060 4980 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Dec 06 03:34:00 crc kubenswrapper[4980]: W1206 03:34:00.981065 4980 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Dec 06 03:34:00 crc kubenswrapper[4980]: W1206 03:34:00.981070 4980 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Dec 06 03:34:00 crc kubenswrapper[4980]: W1206 03:34:00.981075 4980 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Dec 06 03:34:00 crc kubenswrapper[4980]: W1206 03:34:00.981081 4980 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Dec 06 03:34:00 crc kubenswrapper[4980]: W1206 03:34:00.981086 4980 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Dec 06 03:34:00 crc kubenswrapper[4980]: W1206 03:34:00.981091 4980 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Dec 06 03:34:00 crc kubenswrapper[4980]: W1206 03:34:00.981096 4980 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Dec 06 03:34:00 crc kubenswrapper[4980]: W1206 03:34:00.981110 4980 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Dec 06 03:34:00 crc kubenswrapper[4980]: I1206 03:34:00.981118 4980 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Dec 06 03:34:00 crc kubenswrapper[4980]: I1206 03:34:00.981712 4980 server.go:940] "Client rotation is on, will bootstrap in background" Dec 06 03:34:00 crc kubenswrapper[4980]: I1206 03:34:00.985931 4980 bootstrap.go:85] "Current kubeconfig file contents are still valid, no bootstrap necessary" Dec 06 03:34:00 crc kubenswrapper[4980]: I1206 03:34:00.986094 4980 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-client-current.pem". Dec 06 03:34:00 crc kubenswrapper[4980]: I1206 03:34:00.986801 4980 server.go:997] "Starting client certificate rotation" Dec 06 03:34:00 crc kubenswrapper[4980]: I1206 03:34:00.986841 4980 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate rotation is enabled Dec 06 03:34:00 crc kubenswrapper[4980]: I1206 03:34:00.987165 4980 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate expiration is 2026-02-24 05:52:08 +0000 UTC, rotation deadline is 2025-12-11 19:53:44.464459552 +0000 UTC Dec 06 03:34:00 crc kubenswrapper[4980]: I1206 03:34:00.987322 4980 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Waiting 136h19m43.477143171s for next certificate rotation Dec 06 03:34:00 crc kubenswrapper[4980]: I1206 03:34:00.994647 4980 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Dec 06 03:34:00 crc kubenswrapper[4980]: I1206 03:34:00.996407 4980 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.040069 4980 log.go:25] "Validated CRI v1 runtime API" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.061508 4980 log.go:25] "Validated CRI v1 image API" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.064250 4980 server.go:1437] "Using cgroup driver setting received from the CRI runtime" cgroupDriver="systemd" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.068547 4980 fs.go:133] Filesystem UUIDs: map[0b076daa-c26a-46d2-b3a6-72a8dbc6e257:/dev/vda4 2025-12-06-03-30-09-00:/dev/sr0 7B77-95E7:/dev/vda2 de0497b0-db1b-465a-b278-03db02455c71:/dev/vda3] Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.068597 4980 fs.go:134] Filesystem partitions: map[/dev/shm:{mountpoint:/dev/shm major:0 minor:22 fsType:tmpfs blockSize:0} /dev/vda3:{mountpoint:/boot major:252 minor:3 fsType:ext4 blockSize:0} /dev/vda4:{mountpoint:/var major:252 minor:4 fsType:xfs blockSize:0} /run:{mountpoint:/run major:0 minor:24 fsType:tmpfs blockSize:0} /run/user/1000:{mountpoint:/run/user/1000 major:0 minor:42 fsType:tmpfs blockSize:0} /tmp:{mountpoint:/tmp major:0 minor:30 fsType:tmpfs blockSize:0} /var/lib/etcd:{mountpoint:/var/lib/etcd major:0 minor:43 fsType:tmpfs blockSize:0}] Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.085444 4980 manager.go:217] Machine: {Timestamp:2025-12-06 03:34:01.083665209 +0000 UTC m=+0.328416500 CPUVendorID:AuthenticAMD NumCores:12 NumPhysicalCores:1 NumSockets:12 CpuFrequency:2800000 MemoryCapacity:33654132736 SwapCapacity:0 MemoryByType:map[] NVMInfo:{MemoryModeCapacity:0 AppDirectModeCapacity:0 AvgPowerBudget:0} HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] MachineID:21801e6708c44f15b81395eb736a7cec SystemUUID:a5b493aa-69de-4688-8ff6-1e5f49b7f014 BootID:9b4ed4f8-ed8d-447b-9e1f-13defbdcf0f7 Filesystems:[{Device:/var/lib/etcd DeviceMajor:0 DeviceMinor:43 Capacity:1073741824 Type:vfs Inodes:4108170 HasInodes:true} {Device:/dev/shm DeviceMajor:0 DeviceMinor:22 Capacity:16827064320 Type:vfs Inodes:4108170 HasInodes:true} {Device:/run DeviceMajor:0 DeviceMinor:24 Capacity:6730829824 Type:vfs Inodes:819200 HasInodes:true} {Device:/dev/vda4 DeviceMajor:252 DeviceMinor:4 Capacity:85292941312 Type:vfs Inodes:41679680 HasInodes:true} {Device:/tmp DeviceMajor:0 DeviceMinor:30 Capacity:16827068416 Type:vfs Inodes:1048576 HasInodes:true} {Device:/dev/vda3 DeviceMajor:252 DeviceMinor:3 Capacity:366869504 Type:vfs Inodes:98304 HasInodes:true} {Device:/run/user/1000 DeviceMajor:0 DeviceMinor:42 Capacity:3365412864 Type:vfs Inodes:821634 HasInodes:true}] DiskMap:map[252:0:{Name:vda Major:252 Minor:0 Size:214748364800 Scheduler:none}] NetworkDevices:[{Name:br-ex MacAddress:fa:16:3e:48:7c:2a Speed:0 Mtu:1500} {Name:br-int MacAddress:d6:39:55:2e:22:71 Speed:0 Mtu:1400} {Name:ens3 MacAddress:fa:16:3e:48:7c:2a Speed:-1 Mtu:1500} {Name:ens7 MacAddress:fa:16:3e:81:9b:a1 Speed:-1 Mtu:1500} {Name:ens7.20 MacAddress:52:54:00:a6:3b:67 Speed:-1 Mtu:1496} {Name:ens7.21 MacAddress:52:54:00:f3:de:9a Speed:-1 Mtu:1496} {Name:ens7.22 MacAddress:52:54:00:66:9f:6e Speed:-1 Mtu:1496} {Name:eth10 MacAddress:f2:ae:e8:80:c2:81 Speed:0 Mtu:1500} {Name:ovn-k8s-mp0 MacAddress:0a:58:0a:d9:00:02 Speed:0 Mtu:1400} {Name:ovs-system MacAddress:62:b1:b5:e3:14:dd Speed:0 Mtu:1500}] Topology:[{Id:0 Memory:33654132736 HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] Cores:[{Id:0 Threads:[0] Caches:[{Id:0 Size:32768 Type:Data Level:1} {Id:0 Size:32768 Type:Instruction Level:1} {Id:0 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:0 Size:16777216 Type:Unified Level:3}] SocketID:0 BookID: DrawerID:} {Id:0 Threads:[1] Caches:[{Id:1 Size:32768 Type:Data Level:1} {Id:1 Size:32768 Type:Instruction Level:1} {Id:1 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:1 Size:16777216 Type:Unified Level:3}] SocketID:1 BookID: DrawerID:} {Id:0 Threads:[10] Caches:[{Id:10 Size:32768 Type:Data Level:1} {Id:10 Size:32768 Type:Instruction Level:1} {Id:10 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:10 Size:16777216 Type:Unified Level:3}] SocketID:10 BookID: DrawerID:} {Id:0 Threads:[11] Caches:[{Id:11 Size:32768 Type:Data Level:1} {Id:11 Size:32768 Type:Instruction Level:1} {Id:11 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:11 Size:16777216 Type:Unified Level:3}] SocketID:11 BookID: DrawerID:} {Id:0 Threads:[2] Caches:[{Id:2 Size:32768 Type:Data Level:1} {Id:2 Size:32768 Type:Instruction Level:1} {Id:2 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:2 Size:16777216 Type:Unified Level:3}] SocketID:2 BookID: DrawerID:} {Id:0 Threads:[3] Caches:[{Id:3 Size:32768 Type:Data Level:1} {Id:3 Size:32768 Type:Instruction Level:1} {Id:3 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:3 Size:16777216 Type:Unified Level:3}] SocketID:3 BookID: DrawerID:} {Id:0 Threads:[4] Caches:[{Id:4 Size:32768 Type:Data Level:1} {Id:4 Size:32768 Type:Instruction Level:1} {Id:4 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:4 Size:16777216 Type:Unified Level:3}] SocketID:4 BookID: DrawerID:} {Id:0 Threads:[5] Caches:[{Id:5 Size:32768 Type:Data Level:1} {Id:5 Size:32768 Type:Instruction Level:1} {Id:5 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:5 Size:16777216 Type:Unified Level:3}] SocketID:5 BookID: DrawerID:} {Id:0 Threads:[6] Caches:[{Id:6 Size:32768 Type:Data Level:1} {Id:6 Size:32768 Type:Instruction Level:1} {Id:6 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:6 Size:16777216 Type:Unified Level:3}] SocketID:6 BookID: DrawerID:} {Id:0 Threads:[7] Caches:[{Id:7 Size:32768 Type:Data Level:1} {Id:7 Size:32768 Type:Instruction Level:1} {Id:7 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:7 Size:16777216 Type:Unified Level:3}] SocketID:7 BookID: DrawerID:} {Id:0 Threads:[8] Caches:[{Id:8 Size:32768 Type:Data Level:1} {Id:8 Size:32768 Type:Instruction Level:1} {Id:8 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:8 Size:16777216 Type:Unified Level:3}] SocketID:8 BookID: DrawerID:} {Id:0 Threads:[9] Caches:[{Id:9 Size:32768 Type:Data Level:1} {Id:9 Size:32768 Type:Instruction Level:1} {Id:9 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:9 Size:16777216 Type:Unified Level:3}] SocketID:9 BookID: DrawerID:}] Caches:[] Distances:[10]}] CloudProvider:Unknown InstanceType:Unknown InstanceID:None} Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.085775 4980 manager_no_libpfm.go:29] cAdvisor is build without cgo and/or libpfm support. Perf event counters are not available. Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.085999 4980 manager.go:233] Version: {KernelVersion:5.14.0-427.50.2.el9_4.x86_64 ContainerOsVersion:Red Hat Enterprise Linux CoreOS 418.94.202502100215-0 DockerVersion: DockerAPIVersion: CadvisorVersion: CadvisorRevision:} Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.086681 4980 swap_util.go:113] "Swap is on" /proc/swaps contents="Filename\t\t\t\tType\t\tSize\t\tUsed\t\tPriority" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.086884 4980 container_manager_linux.go:267] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.086932 4980 container_manager_linux.go:272] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"crc","RuntimeCgroupsName":"/system.slice/crio.service","SystemCgroupsName":"/system.slice","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":true,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":{"cpu":"200m","ephemeral-storage":"350Mi","memory":"350Mi"},"HardEvictionThresholds":[{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":4096,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.087185 4980 topology_manager.go:138] "Creating topology manager with none policy" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.087198 4980 container_manager_linux.go:303] "Creating device plugin manager" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.087428 4980 manager.go:142] "Creating Device Plugin manager" path="/var/lib/kubelet/device-plugins/kubelet.sock" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.087482 4980 server.go:66] "Creating device plugin registration server" version="v1beta1" socket="/var/lib/kubelet/device-plugins/kubelet.sock" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.087769 4980 state_mem.go:36] "Initialized new in-memory state store" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.087898 4980 server.go:1245] "Using root directory" path="/var/lib/kubelet" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.089719 4980 kubelet.go:418] "Attempting to sync node with API server" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.089749 4980 kubelet.go:313] "Adding static pod path" path="/etc/kubernetes/manifests" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.089778 4980 file.go:69] "Watching path" path="/etc/kubernetes/manifests" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.089801 4980 kubelet.go:324] "Adding apiserver pod source" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.089824 4980 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.092940 4980 kuberuntime_manager.go:262] "Container runtime initialized" containerRuntime="cri-o" version="1.31.5-4.rhaos4.18.gitdad78d5.el9" apiVersion="v1" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.093637 4980 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-server-current.pem". Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.094988 4980 kubelet.go:854] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Dec 06 03:34:01 crc kubenswrapper[4980]: W1206 03:34:01.095440 4980 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.184:6443: connect: connection refused Dec 06 03:34:01 crc kubenswrapper[4980]: W1206 03:34:01.095590 4980 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.184:6443: connect: connection refused Dec 06 03:34:01 crc kubenswrapper[4980]: E1206 03:34:01.095671 4980 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.184:6443: connect: connection refused" logger="UnhandledError" Dec 06 03:34:01 crc kubenswrapper[4980]: E1206 03:34:01.095745 4980 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.184:6443: connect: connection refused" logger="UnhandledError" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.096112 4980 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/portworx-volume" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.096163 4980 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/empty-dir" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.096185 4980 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/git-repo" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.096210 4980 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/host-path" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.096238 4980 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/nfs" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.096257 4980 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/secret" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.096285 4980 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/iscsi" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.096315 4980 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/downward-api" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.096337 4980 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/fc" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.096352 4980 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/configmap" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.096380 4980 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/projected" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.096395 4980 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/local-volume" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.096945 4980 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/csi" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.099017 4980 server.go:1280] "Started kubelet" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.099071 4980 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.184:6443: connect: connection refused Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.100763 4980 server.go:163] "Starting to listen" address="0.0.0.0" port=10250 Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.100779 4980 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Dec 06 03:34:01 crc systemd[1]: Started Kubernetes Kubelet. Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.106082 4980 server.go:236] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.112569 4980 server.go:460] "Adding debug handlers to kubelet server" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.112944 4980 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate rotation is enabled Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.113025 4980 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.113485 4980 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-07 21:28:26.459194383 +0000 UTC Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.113642 4980 certificate_manager.go:356] kubernetes.io/kubelet-serving: Waiting 785h54m25.345560088s for next certificate rotation Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.114017 4980 volume_manager.go:287] "The desired_state_of_world populator starts" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.114074 4980 volume_manager.go:289] "Starting Kubelet Volume Manager" Dec 06 03:34:01 crc kubenswrapper[4980]: E1206 03:34:01.114016 4980 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.114177 4980 desired_state_of_world_populator.go:146] "Desired state populator starts to run" Dec 06 03:34:01 crc kubenswrapper[4980]: E1206 03:34:01.112671 4980 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": dial tcp 38.102.83.184:6443: connect: connection refused" event="&Event{ObjectMeta:{crc.187e82e8f918085d default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-06 03:34:01.098938461 +0000 UTC m=+0.343689822,LastTimestamp:2025-12-06 03:34:01.098938461 +0000 UTC m=+0.343689822,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Dec 06 03:34:01 crc kubenswrapper[4980]: E1206 03:34:01.114739 4980 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.184:6443: connect: connection refused" interval="200ms" Dec 06 03:34:01 crc kubenswrapper[4980]: W1206 03:34:01.114784 4980 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.184:6443: connect: connection refused Dec 06 03:34:01 crc kubenswrapper[4980]: E1206 03:34:01.115141 4980 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.184:6443: connect: connection refused" logger="UnhandledError" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.117182 4980 factory.go:55] Registering systemd factory Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.117474 4980 factory.go:221] Registration of the systemd container factory successfully Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.121501 4980 factory.go:153] Registering CRI-O factory Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.121866 4980 factory.go:221] Registration of the crio container factory successfully Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.122076 4980 factory.go:219] Registration of the containerd container factory failed: unable to create containerd client: containerd: cannot unix dial containerd api service: dial unix /run/containerd/containerd.sock: connect: no such file or directory Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.122283 4980 factory.go:103] Registering Raw factory Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.122379 4980 manager.go:1196] Started watching for new ooms in manager Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.123499 4980 manager.go:319] Starting recovery of all containers Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.133363 4980 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" seLinuxMountContext="" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.138425 4980 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" seLinuxMountContext="" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.138502 4980 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert" seLinuxMountContext="" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.138549 4980 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" seLinuxMountContext="" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.138569 4980 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" seLinuxMountContext="" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.138592 4980 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" seLinuxMountContext="" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.138614 4980 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" seLinuxMountContext="" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.138635 4980 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" seLinuxMountContext="" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.138697 4980 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" seLinuxMountContext="" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.138715 4980 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" seLinuxMountContext="" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.138741 4980 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" seLinuxMountContext="" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.138762 4980 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides" seLinuxMountContext="" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.138781 4980 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" seLinuxMountContext="" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.138811 4980 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" seLinuxMountContext="" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.138837 4980 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" seLinuxMountContext="" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.138860 4980 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" seLinuxMountContext="" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.138875 4980 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" seLinuxMountContext="" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.138895 4980 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" seLinuxMountContext="" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.138910 4980 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" seLinuxMountContext="" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.138925 4980 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" seLinuxMountContext="" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.138942 4980 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" seLinuxMountContext="" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.138959 4980 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" seLinuxMountContext="" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.138977 4980 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" seLinuxMountContext="" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.138991 4980 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" seLinuxMountContext="" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.139010 4980 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" seLinuxMountContext="" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.139028 4980 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" seLinuxMountContext="" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.141867 4980 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" seLinuxMountContext="" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.141905 4980 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" seLinuxMountContext="" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.141922 4980 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" seLinuxMountContext="" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.141941 4980 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls" seLinuxMountContext="" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.141956 4980 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" seLinuxMountContext="" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.141972 4980 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" seLinuxMountContext="" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.141985 4980 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" volumeName="kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" seLinuxMountContext="" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.142000 4980 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" seLinuxMountContext="" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.142018 4980 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" seLinuxMountContext="" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.142034 4980 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" seLinuxMountContext="" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.142045 4980 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" seLinuxMountContext="" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.142061 4980 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" seLinuxMountContext="" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.142076 4980 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" seLinuxMountContext="" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.142091 4980 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" seLinuxMountContext="" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.142107 4980 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" seLinuxMountContext="" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.142131 4980 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" seLinuxMountContext="" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.142147 4980 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" seLinuxMountContext="" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.142162 4980 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" seLinuxMountContext="" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.142177 4980 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" seLinuxMountContext="" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.142218 4980 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" seLinuxMountContext="" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.142234 4980 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" seLinuxMountContext="" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.142248 4980 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" seLinuxMountContext="" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.142293 4980 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="44663579-783b-4372-86d6-acf235a62d72" volumeName="kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" seLinuxMountContext="" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.142310 4980 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" seLinuxMountContext="" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.142331 4980 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" seLinuxMountContext="" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.142347 4980 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf" seLinuxMountContext="" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.142370 4980 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" seLinuxMountContext="" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.142392 4980 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" seLinuxMountContext="" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.142417 4980 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" seLinuxMountContext="" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.142433 4980 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" seLinuxMountContext="" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.142450 4980 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" seLinuxMountContext="" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.142466 4980 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" seLinuxMountContext="" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.142482 4980 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" seLinuxMountContext="" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.142497 4980 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" seLinuxMountContext="" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.142532 4980 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" seLinuxMountContext="" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.142550 4980 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" seLinuxMountContext="" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.142565 4980 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" seLinuxMountContext="" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.142581 4980 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" seLinuxMountContext="" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.142596 4980 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf" seLinuxMountContext="" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.142612 4980 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" seLinuxMountContext="" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.142629 4980 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" seLinuxMountContext="" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.142668 4980 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" seLinuxMountContext="" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.142685 4980 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" seLinuxMountContext="" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.142701 4980 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" seLinuxMountContext="" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.142716 4980 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" seLinuxMountContext="" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.142730 4980 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" volumeName="kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" seLinuxMountContext="" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.142746 4980 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" seLinuxMountContext="" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.142761 4980 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" seLinuxMountContext="" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.142778 4980 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" seLinuxMountContext="" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.142793 4980 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" seLinuxMountContext="" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.142809 4980 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" seLinuxMountContext="" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.142825 4980 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" seLinuxMountContext="" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.142841 4980 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script" seLinuxMountContext="" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.142860 4980 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" seLinuxMountContext="" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.142877 4980 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" seLinuxMountContext="" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.142894 4980 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" seLinuxMountContext="" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.142909 4980 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" seLinuxMountContext="" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.142923 4980 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" seLinuxMountContext="" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.142937 4980 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" seLinuxMountContext="" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.142952 4980 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" seLinuxMountContext="" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.142967 4980 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" seLinuxMountContext="" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.142983 4980 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" seLinuxMountContext="" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.143000 4980 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" seLinuxMountContext="" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.143018 4980 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" seLinuxMountContext="" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.143037 4980 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" seLinuxMountContext="" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.143081 4980 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" seLinuxMountContext="" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.143098 4980 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" seLinuxMountContext="" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.143113 4980 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" seLinuxMountContext="" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.143128 4980 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" seLinuxMountContext="" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.143144 4980 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" seLinuxMountContext="" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.143163 4980 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" seLinuxMountContext="" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.143179 4980 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" seLinuxMountContext="" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.143195 4980 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" seLinuxMountContext="" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.143226 4980 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" seLinuxMountContext="" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.143246 4980 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" seLinuxMountContext="" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.143265 4980 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" seLinuxMountContext="" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.143279 4980 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5" seLinuxMountContext="" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.143295 4980 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" seLinuxMountContext="" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.143379 4980 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" seLinuxMountContext="" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.143399 4980 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" seLinuxMountContext="" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.143416 4980 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" seLinuxMountContext="" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.143439 4980 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" seLinuxMountContext="" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.143458 4980 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" seLinuxMountContext="" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.143478 4980 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" seLinuxMountContext="" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.143493 4980 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" seLinuxMountContext="" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.143528 4980 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" seLinuxMountContext="" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.143546 4980 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb" seLinuxMountContext="" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.143562 4980 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" seLinuxMountContext="" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.143579 4980 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" seLinuxMountContext="" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.143598 4980 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" seLinuxMountContext="" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.143613 4980 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" seLinuxMountContext="" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.143627 4980 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" seLinuxMountContext="" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.143644 4980 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" seLinuxMountContext="" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.143659 4980 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" seLinuxMountContext="" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.143674 4980 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" seLinuxMountContext="" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.143689 4980 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" seLinuxMountContext="" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.143706 4980 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" seLinuxMountContext="" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.143730 4980 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" seLinuxMountContext="" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.143746 4980 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" seLinuxMountContext="" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.143762 4980 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" seLinuxMountContext="" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.143776 4980 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" seLinuxMountContext="" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.143791 4980 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" seLinuxMountContext="" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.143805 4980 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" seLinuxMountContext="" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.143820 4980 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" seLinuxMountContext="" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.143835 4980 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm" seLinuxMountContext="" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.143861 4980 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" seLinuxMountContext="" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.143875 4980 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" seLinuxMountContext="" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.143889 4980 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" seLinuxMountContext="" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.143905 4980 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" seLinuxMountContext="" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.143919 4980 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" seLinuxMountContext="" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.143933 4980 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" seLinuxMountContext="" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.143947 4980 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" seLinuxMountContext="" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.143979 4980 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" seLinuxMountContext="" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.144008 4980 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" seLinuxMountContext="" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.144024 4980 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" seLinuxMountContext="" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.144040 4980 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" seLinuxMountContext="" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.144053 4980 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" seLinuxMountContext="" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.144066 4980 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" seLinuxMountContext="" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.144080 4980 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" seLinuxMountContext="" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.144094 4980 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert" seLinuxMountContext="" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.144107 4980 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" seLinuxMountContext="" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.144145 4980 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" seLinuxMountContext="" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.144159 4980 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" seLinuxMountContext="" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.144172 4980 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" seLinuxMountContext="" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.144187 4980 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" seLinuxMountContext="" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.144215 4980 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49ef4625-1d3a-4a9f-b595-c2433d32326d" volumeName="kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" seLinuxMountContext="" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.144235 4980 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" seLinuxMountContext="" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.144250 4980 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" seLinuxMountContext="" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.144265 4980 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" seLinuxMountContext="" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.144295 4980 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" seLinuxMountContext="" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.144311 4980 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" seLinuxMountContext="" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.144325 4980 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" seLinuxMountContext="" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.144339 4980 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" seLinuxMountContext="" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.144352 4980 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" seLinuxMountContext="" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.144364 4980 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" seLinuxMountContext="" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.144382 4980 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" seLinuxMountContext="" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.144396 4980 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" seLinuxMountContext="" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.144443 4980 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" seLinuxMountContext="" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.144459 4980 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" seLinuxMountContext="" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.144473 4980 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" seLinuxMountContext="" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.144488 4980 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" seLinuxMountContext="" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.144502 4980 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" seLinuxMountContext="" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.144565 4980 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" seLinuxMountContext="" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.144582 4980 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" seLinuxMountContext="" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.144597 4980 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" seLinuxMountContext="" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.144630 4980 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" seLinuxMountContext="" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.144644 4980 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" seLinuxMountContext="" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.144657 4980 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" seLinuxMountContext="" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.144672 4980 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" seLinuxMountContext="" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.144687 4980 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" seLinuxMountContext="" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.144700 4980 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" seLinuxMountContext="" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.144716 4980 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" seLinuxMountContext="" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.150197 4980 reconstruct.go:144] "Volume is marked device as uncertain and added into the actual state" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" deviceMountPath="/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.150334 4980 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" seLinuxMountContext="" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.150360 4980 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" seLinuxMountContext="" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.150379 4980 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" seLinuxMountContext="" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.150396 4980 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" seLinuxMountContext="" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.150415 4980 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" seLinuxMountContext="" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.150436 4980 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" seLinuxMountContext="" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.150452 4980 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3b6479f0-333b-4a96-9adf-2099afdc2447" volumeName="kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr" seLinuxMountContext="" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.150477 4980 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" seLinuxMountContext="" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.150571 4980 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d751cbb-f2e2-430d-9754-c882a5e924a5" volumeName="kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl" seLinuxMountContext="" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.150591 4980 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" seLinuxMountContext="" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.150613 4980 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" seLinuxMountContext="" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.150632 4980 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" seLinuxMountContext="" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.150647 4980 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" seLinuxMountContext="" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.150666 4980 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" seLinuxMountContext="" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.150793 4980 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" seLinuxMountContext="" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.151498 4980 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" seLinuxMountContext="" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.151584 4980 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" seLinuxMountContext="" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.151601 4980 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" seLinuxMountContext="" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.151675 4980 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" seLinuxMountContext="" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.151745 4980 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" seLinuxMountContext="" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.151762 4980 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" seLinuxMountContext="" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.151780 4980 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" seLinuxMountContext="" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.151799 4980 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" seLinuxMountContext="" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.151816 4980 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" seLinuxMountContext="" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.151860 4980 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" seLinuxMountContext="" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.151877 4980 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" seLinuxMountContext="" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.151895 4980 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" seLinuxMountContext="" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.151917 4980 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" seLinuxMountContext="" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.151941 4980 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" seLinuxMountContext="" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.151959 4980 reconstruct.go:97] "Volume reconstruction finished" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.151971 4980 reconciler.go:26] "Reconciler: start to sync state" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.157541 4980 manager.go:324] Recovery completed Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.168749 4980 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.170793 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.170861 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.170877 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.172010 4980 cpu_manager.go:225] "Starting CPU manager" policy="none" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.172029 4980 cpu_manager.go:226] "Reconciling" reconcilePeriod="10s" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.172051 4980 state_mem.go:36] "Initialized new in-memory state store" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.181013 4980 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.181847 4980 policy_none.go:49] "None policy: Start" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.183003 4980 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.183069 4980 status_manager.go:217] "Starting to sync pod status with apiserver" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.183111 4980 kubelet.go:2335] "Starting kubelet main sync loop" Dec 06 03:34:01 crc kubenswrapper[4980]: E1206 03:34:01.183180 4980 kubelet.go:2359] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Dec 06 03:34:01 crc kubenswrapper[4980]: W1206 03:34:01.184055 4980 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.184:6443: connect: connection refused Dec 06 03:34:01 crc kubenswrapper[4980]: E1206 03:34:01.184119 4980 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.184:6443: connect: connection refused" logger="UnhandledError" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.185242 4980 memory_manager.go:170] "Starting memorymanager" policy="None" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.185333 4980 state_mem.go:35] "Initializing new in-memory state store" Dec 06 03:34:01 crc kubenswrapper[4980]: E1206 03:34:01.214584 4980 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.243733 4980 manager.go:334] "Starting Device Plugin manager" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.243810 4980 manager.go:513] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.243825 4980 server.go:79] "Starting device plugin registration server" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.244413 4980 eviction_manager.go:189] "Eviction manager: starting control loop" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.244436 4980 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.245083 4980 plugin_watcher.go:51] "Plugin Watcher Start" path="/var/lib/kubelet/plugins_registry" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.245264 4980 plugin_manager.go:116] "The desired_state_of_world populator (plugin watcher) starts" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.245278 4980 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Dec 06 03:34:01 crc kubenswrapper[4980]: E1206 03:34:01.261175 4980 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.321709 4980 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-crc","openshift-etcd/etcd-crc","openshift-kube-apiserver/kube-apiserver-crc","openshift-kube-controller-manager/kube-controller-manager-crc","openshift-kube-scheduler/openshift-kube-scheduler-crc"] Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.321865 4980 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 03:34:01 crc kubenswrapper[4980]: E1206 03:34:01.322943 4980 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.184:6443: connect: connection refused" interval="400ms" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.324253 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.324301 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.324315 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.324554 4980 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.324835 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.324941 4980 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.325824 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.325885 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.325907 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.326027 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.326064 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.326080 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.326138 4980 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.326558 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.326610 4980 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.327765 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.327781 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.327818 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.327829 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.327848 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.327870 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.328051 4980 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.328569 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.328617 4980 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.329633 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.329666 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.329670 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.329682 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.329729 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.329741 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.330848 4980 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.331451 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.331481 4980 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.332395 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.332419 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.332432 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.332612 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.332664 4980 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.332679 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.332702 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.332713 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.333706 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.333739 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.333752 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.359971 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.360030 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.369690 4980 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.370928 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.370966 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.370976 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.371000 4980 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 06 03:34:01 crc kubenswrapper[4980]: E1206 03:34:01.371492 4980 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.184:6443: connect: connection refused" node="crc" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.461401 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.461506 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.461590 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.461628 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.461660 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.461694 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.461726 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.461757 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.461786 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.461816 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.461844 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.461872 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.461901 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.461933 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.461964 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.462497 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.462553 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.563654 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.563747 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.563796 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.563841 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.563855 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.563935 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.564098 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.564119 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.564496 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.564631 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.564693 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.564628 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.564682 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.564850 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.564755 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.564897 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.564947 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.564992 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.564995 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.564949 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.565090 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.565147 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.565224 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.565250 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.565323 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.565418 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.572106 4980 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.574041 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.574099 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.574115 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.574152 4980 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 06 03:34:01 crc kubenswrapper[4980]: E1206 03:34:01.574557 4980 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.184:6443: connect: connection refused" node="crc" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.663616 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.675721 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.693345 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 06 03:34:01 crc kubenswrapper[4980]: W1206 03:34:01.696766 4980 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd1b160f5dda77d281dd8e69ec8d817f9.slice/crio-893d8354f26f257b4a696f8b3e571d40f40f70d98742541e91e7b543538f474c WatchSource:0}: Error finding container 893d8354f26f257b4a696f8b3e571d40f40f70d98742541e91e7b543538f474c: Status 404 returned error can't find the container with id 893d8354f26f257b4a696f8b3e571d40f40f70d98742541e91e7b543538f474c Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.698984 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 06 03:34:01 crc kubenswrapper[4980]: W1206 03:34:01.701255 4980 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2139d3e2895fc6797b9c76a1b4c9886d.slice/crio-d7b272768f78e26e414d2c61219304ad3ce1780dfc5f027e1018159ab100d297 WatchSource:0}: Error finding container d7b272768f78e26e414d2c61219304ad3ce1780dfc5f027e1018159ab100d297: Status 404 returned error can't find the container with id d7b272768f78e26e414d2c61219304ad3ce1780dfc5f027e1018159ab100d297 Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.703571 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 06 03:34:01 crc kubenswrapper[4980]: W1206 03:34:01.719172 4980 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf4b27818a5e8e43d0dc095d08835c792.slice/crio-eda44fe599b02ac308f86fb88b4cfa8ffbcf5c1e9df6f75c82cda81d1390ccda WatchSource:0}: Error finding container eda44fe599b02ac308f86fb88b4cfa8ffbcf5c1e9df6f75c82cda81d1390ccda: Status 404 returned error can't find the container with id eda44fe599b02ac308f86fb88b4cfa8ffbcf5c1e9df6f75c82cda81d1390ccda Dec 06 03:34:01 crc kubenswrapper[4980]: E1206 03:34:01.724692 4980 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.184:6443: connect: connection refused" interval="800ms" Dec 06 03:34:01 crc kubenswrapper[4980]: W1206 03:34:01.738083 4980 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf614b9022728cf315e60c057852e563e.slice/crio-6b404843d0b4bec66b85afc473ca6d0153172b451bd8458b1149668cc6eaa22e WatchSource:0}: Error finding container 6b404843d0b4bec66b85afc473ca6d0153172b451bd8458b1149668cc6eaa22e: Status 404 returned error can't find the container with id 6b404843d0b4bec66b85afc473ca6d0153172b451bd8458b1149668cc6eaa22e Dec 06 03:34:01 crc kubenswrapper[4980]: W1206 03:34:01.763141 4980 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3dcd261975c3d6b9a6ad6367fd4facd3.slice/crio-c020e546ce2a08134b9dadf62ab0a4548f05caab3e84f411e6b3fdb1ebe7b03c WatchSource:0}: Error finding container c020e546ce2a08134b9dadf62ab0a4548f05caab3e84f411e6b3fdb1ebe7b03c: Status 404 returned error can't find the container with id c020e546ce2a08134b9dadf62ab0a4548f05caab3e84f411e6b3fdb1ebe7b03c Dec 06 03:34:01 crc kubenswrapper[4980]: W1206 03:34:01.927507 4980 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.184:6443: connect: connection refused Dec 06 03:34:01 crc kubenswrapper[4980]: E1206 03:34:01.927691 4980 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.184:6443: connect: connection refused" logger="UnhandledError" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.975447 4980 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.978096 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.978171 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.978195 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:34:01 crc kubenswrapper[4980]: I1206 03:34:01.978242 4980 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 06 03:34:01 crc kubenswrapper[4980]: E1206 03:34:01.978848 4980 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.184:6443: connect: connection refused" node="crc" Dec 06 03:34:01 crc kubenswrapper[4980]: W1206 03:34:01.990506 4980 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.184:6443: connect: connection refused Dec 06 03:34:01 crc kubenswrapper[4980]: E1206 03:34:01.990623 4980 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.184:6443: connect: connection refused" logger="UnhandledError" Dec 06 03:34:02 crc kubenswrapper[4980]: W1206 03:34:02.027348 4980 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.184:6443: connect: connection refused Dec 06 03:34:02 crc kubenswrapper[4980]: E1206 03:34:02.027499 4980 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.184:6443: connect: connection refused" logger="UnhandledError" Dec 06 03:34:02 crc kubenswrapper[4980]: I1206 03:34:02.099753 4980 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.184:6443: connect: connection refused Dec 06 03:34:02 crc kubenswrapper[4980]: I1206 03:34:02.188952 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"eda44fe599b02ac308f86fb88b4cfa8ffbcf5c1e9df6f75c82cda81d1390ccda"} Dec 06 03:34:02 crc kubenswrapper[4980]: I1206 03:34:02.189817 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"d7b272768f78e26e414d2c61219304ad3ce1780dfc5f027e1018159ab100d297"} Dec 06 03:34:02 crc kubenswrapper[4980]: I1206 03:34:02.190738 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"893d8354f26f257b4a696f8b3e571d40f40f70d98742541e91e7b543538f474c"} Dec 06 03:34:02 crc kubenswrapper[4980]: I1206 03:34:02.191497 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"c020e546ce2a08134b9dadf62ab0a4548f05caab3e84f411e6b3fdb1ebe7b03c"} Dec 06 03:34:02 crc kubenswrapper[4980]: I1206 03:34:02.192199 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"6b404843d0b4bec66b85afc473ca6d0153172b451bd8458b1149668cc6eaa22e"} Dec 06 03:34:02 crc kubenswrapper[4980]: W1206 03:34:02.384159 4980 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.184:6443: connect: connection refused Dec 06 03:34:02 crc kubenswrapper[4980]: E1206 03:34:02.384293 4980 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.184:6443: connect: connection refused" logger="UnhandledError" Dec 06 03:34:02 crc kubenswrapper[4980]: E1206 03:34:02.526317 4980 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.184:6443: connect: connection refused" interval="1.6s" Dec 06 03:34:02 crc kubenswrapper[4980]: I1206 03:34:02.779096 4980 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 03:34:02 crc kubenswrapper[4980]: I1206 03:34:02.781260 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:34:02 crc kubenswrapper[4980]: I1206 03:34:02.781316 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:34:02 crc kubenswrapper[4980]: I1206 03:34:02.781325 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:34:02 crc kubenswrapper[4980]: I1206 03:34:02.781357 4980 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 06 03:34:02 crc kubenswrapper[4980]: E1206 03:34:02.781896 4980 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.184:6443: connect: connection refused" node="crc" Dec 06 03:34:03 crc kubenswrapper[4980]: I1206 03:34:03.101024 4980 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.184:6443: connect: connection refused Dec 06 03:34:03 crc kubenswrapper[4980]: I1206 03:34:03.198432 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"32af3427e872a70cb3c2357116f04cabc8d3d9b69e7306b2e0de2ae22f9469d4"} Dec 06 03:34:03 crc kubenswrapper[4980]: I1206 03:34:03.198496 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"b543aa4ee03d5a14909b67d876c5675c02e1df2a89a3c9ac8d3fb587fea53d8f"} Dec 06 03:34:03 crc kubenswrapper[4980]: I1206 03:34:03.198542 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"725e84f6bdfb722a8ce69158d03ea4c66a9f087642d55cb9651e003c81be0c8b"} Dec 06 03:34:03 crc kubenswrapper[4980]: I1206 03:34:03.198560 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"71f7b492ea07cdf7dc65fc48184ec1e64914eb1811bbfdddc0fb14c8ad6cb296"} Dec 06 03:34:03 crc kubenswrapper[4980]: I1206 03:34:03.198599 4980 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 03:34:03 crc kubenswrapper[4980]: I1206 03:34:03.200129 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:34:03 crc kubenswrapper[4980]: I1206 03:34:03.200182 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:34:03 crc kubenswrapper[4980]: I1206 03:34:03.200203 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:34:03 crc kubenswrapper[4980]: I1206 03:34:03.201378 4980 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="2cba89f2f1bed96d59f183628d900d9b72c91eb472e273c8bed0babcab485910" exitCode=0 Dec 06 03:34:03 crc kubenswrapper[4980]: I1206 03:34:03.201526 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"2cba89f2f1bed96d59f183628d900d9b72c91eb472e273c8bed0babcab485910"} Dec 06 03:34:03 crc kubenswrapper[4980]: I1206 03:34:03.201555 4980 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 03:34:03 crc kubenswrapper[4980]: I1206 03:34:03.202920 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:34:03 crc kubenswrapper[4980]: I1206 03:34:03.202957 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:34:03 crc kubenswrapper[4980]: I1206 03:34:03.202972 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:34:03 crc kubenswrapper[4980]: I1206 03:34:03.205096 4980 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 03:34:03 crc kubenswrapper[4980]: I1206 03:34:03.206007 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:34:03 crc kubenswrapper[4980]: I1206 03:34:03.206035 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:34:03 crc kubenswrapper[4980]: I1206 03:34:03.206047 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:34:03 crc kubenswrapper[4980]: I1206 03:34:03.206160 4980 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="27cf07cd96893900da4afecd0d56fd9269bc60c1d23246b4685fcd8756357c19" exitCode=0 Dec 06 03:34:03 crc kubenswrapper[4980]: I1206 03:34:03.206281 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"27cf07cd96893900da4afecd0d56fd9269bc60c1d23246b4685fcd8756357c19"} Dec 06 03:34:03 crc kubenswrapper[4980]: I1206 03:34:03.206456 4980 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 03:34:03 crc kubenswrapper[4980]: I1206 03:34:03.207804 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:34:03 crc kubenswrapper[4980]: I1206 03:34:03.207831 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:34:03 crc kubenswrapper[4980]: I1206 03:34:03.207844 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:34:03 crc kubenswrapper[4980]: I1206 03:34:03.209158 4980 generic.go:334] "Generic (PLEG): container finished" podID="d1b160f5dda77d281dd8e69ec8d817f9" containerID="3914c0500d985e18f774363c2213ecde31d60abac5937dc794e72fcfe357857d" exitCode=0 Dec 06 03:34:03 crc kubenswrapper[4980]: I1206 03:34:03.209223 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerDied","Data":"3914c0500d985e18f774363c2213ecde31d60abac5937dc794e72fcfe357857d"} Dec 06 03:34:03 crc kubenswrapper[4980]: I1206 03:34:03.209314 4980 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 03:34:03 crc kubenswrapper[4980]: I1206 03:34:03.210025 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:34:03 crc kubenswrapper[4980]: I1206 03:34:03.210055 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:34:03 crc kubenswrapper[4980]: I1206 03:34:03.210067 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:34:03 crc kubenswrapper[4980]: I1206 03:34:03.214543 4980 generic.go:334] "Generic (PLEG): container finished" podID="3dcd261975c3d6b9a6ad6367fd4facd3" containerID="1f596bfd5321ca225f4be03d4cf229c31a52def9731c0e177ab3a0bc010380e1" exitCode=0 Dec 06 03:34:03 crc kubenswrapper[4980]: I1206 03:34:03.214577 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerDied","Data":"1f596bfd5321ca225f4be03d4cf229c31a52def9731c0e177ab3a0bc010380e1"} Dec 06 03:34:03 crc kubenswrapper[4980]: I1206 03:34:03.214683 4980 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 03:34:03 crc kubenswrapper[4980]: I1206 03:34:03.215637 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:34:03 crc kubenswrapper[4980]: I1206 03:34:03.215662 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:34:03 crc kubenswrapper[4980]: I1206 03:34:03.215675 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:34:03 crc kubenswrapper[4980]: I1206 03:34:03.566218 4980 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 06 03:34:03 crc kubenswrapper[4980]: E1206 03:34:03.713960 4980 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": dial tcp 38.102.83.184:6443: connect: connection refused" event="&Event{ObjectMeta:{crc.187e82e8f918085d default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-06 03:34:01.098938461 +0000 UTC m=+0.343689822,LastTimestamp:2025-12-06 03:34:01.098938461 +0000 UTC m=+0.343689822,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Dec 06 03:34:04 crc kubenswrapper[4980]: W1206 03:34:04.038272 4980 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.184:6443: connect: connection refused Dec 06 03:34:04 crc kubenswrapper[4980]: E1206 03:34:04.038474 4980 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.184:6443: connect: connection refused" logger="UnhandledError" Dec 06 03:34:04 crc kubenswrapper[4980]: I1206 03:34:04.101164 4980 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.184:6443: connect: connection refused Dec 06 03:34:04 crc kubenswrapper[4980]: E1206 03:34:04.134491 4980 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.184:6443: connect: connection refused" interval="3.2s" Dec 06 03:34:04 crc kubenswrapper[4980]: I1206 03:34:04.245859 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"206f6f19fe3197f5d6798cc86c5473bf2decb02a2860ee7ffb5efb5f60952e8c"} Dec 06 03:34:04 crc kubenswrapper[4980]: I1206 03:34:04.245920 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"c28b6895b14e73f4f3eb7a8ce93076c05092ba9b58ea0e3e879c65bcfa227a21"} Dec 06 03:34:04 crc kubenswrapper[4980]: I1206 03:34:04.248414 4980 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="c49feb24566a8284378670fa0390710d38c237acc94ffb654f90f3fa4626e4c8" exitCode=0 Dec 06 03:34:04 crc kubenswrapper[4980]: I1206 03:34:04.248534 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"c49feb24566a8284378670fa0390710d38c237acc94ffb654f90f3fa4626e4c8"} Dec 06 03:34:04 crc kubenswrapper[4980]: I1206 03:34:04.248544 4980 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 03:34:04 crc kubenswrapper[4980]: I1206 03:34:04.249865 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:34:04 crc kubenswrapper[4980]: I1206 03:34:04.249923 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:34:04 crc kubenswrapper[4980]: I1206 03:34:04.249941 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:34:04 crc kubenswrapper[4980]: I1206 03:34:04.251193 4980 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 03:34:04 crc kubenswrapper[4980]: I1206 03:34:04.251175 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"7d2f4b65b063604aa44fe05821c7aa6eed2989422f352b408423cf74c2390ae8"} Dec 06 03:34:04 crc kubenswrapper[4980]: I1206 03:34:04.252114 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:34:04 crc kubenswrapper[4980]: I1206 03:34:04.252156 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:34:04 crc kubenswrapper[4980]: I1206 03:34:04.252176 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:34:04 crc kubenswrapper[4980]: I1206 03:34:04.256033 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"173903f3554d7ac9158493498a016d618d92f46da1500408834b3c24249cf117"} Dec 06 03:34:04 crc kubenswrapper[4980]: I1206 03:34:04.256077 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"ae66052702fcfd94f0123f4fbbe9efed37eb04609269a2bc951f95669de16601"} Dec 06 03:34:04 crc kubenswrapper[4980]: I1206 03:34:04.256099 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"5beac3fd93b448506c0a110d70ef5faf42efa18b16abe6b3d4eb7b9c4233ad93"} Dec 06 03:34:04 crc kubenswrapper[4980]: I1206 03:34:04.256148 4980 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 03:34:04 crc kubenswrapper[4980]: I1206 03:34:04.256184 4980 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 03:34:04 crc kubenswrapper[4980]: I1206 03:34:04.257821 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:34:04 crc kubenswrapper[4980]: I1206 03:34:04.257856 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:34:04 crc kubenswrapper[4980]: I1206 03:34:04.257870 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:34:04 crc kubenswrapper[4980]: I1206 03:34:04.257956 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:34:04 crc kubenswrapper[4980]: I1206 03:34:04.257984 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:34:04 crc kubenswrapper[4980]: I1206 03:34:04.257996 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:34:04 crc kubenswrapper[4980]: I1206 03:34:04.382960 4980 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 03:34:04 crc kubenswrapper[4980]: I1206 03:34:04.399194 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:34:04 crc kubenswrapper[4980]: I1206 03:34:04.399278 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:34:04 crc kubenswrapper[4980]: I1206 03:34:04.399289 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:34:04 crc kubenswrapper[4980]: I1206 03:34:04.399341 4980 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 06 03:34:04 crc kubenswrapper[4980]: E1206 03:34:04.400171 4980 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.184:6443: connect: connection refused" node="crc" Dec 06 03:34:04 crc kubenswrapper[4980]: W1206 03:34:04.501504 4980 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.184:6443: connect: connection refused Dec 06 03:34:04 crc kubenswrapper[4980]: E1206 03:34:04.501935 4980 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.184:6443: connect: connection refused" logger="UnhandledError" Dec 06 03:34:04 crc kubenswrapper[4980]: W1206 03:34:04.665972 4980 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.184:6443: connect: connection refused Dec 06 03:34:04 crc kubenswrapper[4980]: E1206 03:34:04.666117 4980 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.184:6443: connect: connection refused" logger="UnhandledError" Dec 06 03:34:05 crc kubenswrapper[4980]: I1206 03:34:05.262388 4980 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="bb74d1c3d752a379fa394c450217badb45511fbcafad62f7b618131cab94b29e" exitCode=0 Dec 06 03:34:05 crc kubenswrapper[4980]: I1206 03:34:05.262452 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"bb74d1c3d752a379fa394c450217badb45511fbcafad62f7b618131cab94b29e"} Dec 06 03:34:05 crc kubenswrapper[4980]: I1206 03:34:05.262697 4980 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 03:34:05 crc kubenswrapper[4980]: I1206 03:34:05.264177 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:34:05 crc kubenswrapper[4980]: I1206 03:34:05.264216 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:34:05 crc kubenswrapper[4980]: I1206 03:34:05.264230 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:34:05 crc kubenswrapper[4980]: I1206 03:34:05.266629 4980 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 03:34:05 crc kubenswrapper[4980]: I1206 03:34:05.267244 4980 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 03:34:05 crc kubenswrapper[4980]: I1206 03:34:05.267573 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"773a0bc2aab73ffc2717e987543defc9b786cc22300aa54b4ba4f3454504e331"} Dec 06 03:34:05 crc kubenswrapper[4980]: I1206 03:34:05.267609 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"d40e0e7446f8416bf9da94fc53c01be937f9a51a11eb101157d2ddd4597720d2"} Dec 06 03:34:05 crc kubenswrapper[4980]: I1206 03:34:05.267622 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"76cff8a2d652df38b04eefc053832baf4cbb5db778ff5e525b18c2825141b118"} Dec 06 03:34:05 crc kubenswrapper[4980]: I1206 03:34:05.267671 4980 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 03:34:05 crc kubenswrapper[4980]: I1206 03:34:05.268018 4980 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 06 03:34:05 crc kubenswrapper[4980]: I1206 03:34:05.268049 4980 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 03:34:05 crc kubenswrapper[4980]: I1206 03:34:05.269043 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:34:05 crc kubenswrapper[4980]: I1206 03:34:05.269061 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:34:05 crc kubenswrapper[4980]: I1206 03:34:05.269070 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:34:05 crc kubenswrapper[4980]: I1206 03:34:05.269069 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:34:05 crc kubenswrapper[4980]: I1206 03:34:05.269106 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:34:05 crc kubenswrapper[4980]: I1206 03:34:05.269127 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:34:05 crc kubenswrapper[4980]: I1206 03:34:05.269226 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:34:05 crc kubenswrapper[4980]: I1206 03:34:05.269243 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:34:05 crc kubenswrapper[4980]: I1206 03:34:05.269260 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:34:05 crc kubenswrapper[4980]: I1206 03:34:05.269342 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:34:05 crc kubenswrapper[4980]: I1206 03:34:05.269362 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:34:05 crc kubenswrapper[4980]: I1206 03:34:05.269380 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:34:05 crc kubenswrapper[4980]: I1206 03:34:05.646736 4980 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 06 03:34:05 crc kubenswrapper[4980]: I1206 03:34:05.991023 4980 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 06 03:34:06 crc kubenswrapper[4980]: I1206 03:34:06.273293 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"8c22e7dd8886c2dfbb742900ed86ec7b0c2edf9a1363e74f6666961d928f06c2"} Dec 06 03:34:06 crc kubenswrapper[4980]: I1206 03:34:06.273353 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"2f7fcc44b42bcbb55509048e3a566bb557333c4a83e33d7b0fba8eb514317909"} Dec 06 03:34:06 crc kubenswrapper[4980]: I1206 03:34:06.273371 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"77bc12768b018625ba11a6bfd66688f926c4b7f3bc1d0e41c6551a18b485ecf3"} Dec 06 03:34:06 crc kubenswrapper[4980]: I1206 03:34:06.273423 4980 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 03:34:06 crc kubenswrapper[4980]: I1206 03:34:06.273455 4980 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 06 03:34:06 crc kubenswrapper[4980]: I1206 03:34:06.273557 4980 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 03:34:06 crc kubenswrapper[4980]: I1206 03:34:06.276911 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:34:06 crc kubenswrapper[4980]: I1206 03:34:06.277556 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:34:06 crc kubenswrapper[4980]: I1206 03:34:06.277588 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:34:06 crc kubenswrapper[4980]: I1206 03:34:06.276954 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:34:06 crc kubenswrapper[4980]: I1206 03:34:06.277712 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:34:06 crc kubenswrapper[4980]: I1206 03:34:06.277730 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:34:07 crc kubenswrapper[4980]: I1206 03:34:07.284656 4980 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 06 03:34:07 crc kubenswrapper[4980]: I1206 03:34:07.284719 4980 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 03:34:07 crc kubenswrapper[4980]: I1206 03:34:07.284719 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"1ee8e936dfc39361d6f99460910591ca8c901a20f0d274bbb6aaebfa8a00ee06"} Dec 06 03:34:07 crc kubenswrapper[4980]: I1206 03:34:07.284831 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"473e792b2d0a9644854ee0581c6120fc9c989bc342899ef72b1d056080fd4694"} Dec 06 03:34:07 crc kubenswrapper[4980]: I1206 03:34:07.284970 4980 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 03:34:07 crc kubenswrapper[4980]: I1206 03:34:07.286894 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:34:07 crc kubenswrapper[4980]: I1206 03:34:07.286939 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:34:07 crc kubenswrapper[4980]: I1206 03:34:07.286957 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:34:07 crc kubenswrapper[4980]: I1206 03:34:07.287158 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:34:07 crc kubenswrapper[4980]: I1206 03:34:07.287201 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:34:07 crc kubenswrapper[4980]: I1206 03:34:07.287214 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:34:07 crc kubenswrapper[4980]: I1206 03:34:07.560305 4980 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 06 03:34:07 crc kubenswrapper[4980]: I1206 03:34:07.600629 4980 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 03:34:07 crc kubenswrapper[4980]: I1206 03:34:07.603094 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:34:07 crc kubenswrapper[4980]: I1206 03:34:07.603159 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:34:07 crc kubenswrapper[4980]: I1206 03:34:07.603185 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:34:07 crc kubenswrapper[4980]: I1206 03:34:07.603227 4980 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 06 03:34:08 crc kubenswrapper[4980]: I1206 03:34:08.108462 4980 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 06 03:34:08 crc kubenswrapper[4980]: I1206 03:34:08.115590 4980 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-etcd/etcd-crc" Dec 06 03:34:08 crc kubenswrapper[4980]: I1206 03:34:08.251085 4980 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 06 03:34:08 crc kubenswrapper[4980]: I1206 03:34:08.251427 4980 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 03:34:08 crc kubenswrapper[4980]: I1206 03:34:08.253205 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:34:08 crc kubenswrapper[4980]: I1206 03:34:08.253261 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:34:08 crc kubenswrapper[4980]: I1206 03:34:08.253286 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:34:08 crc kubenswrapper[4980]: I1206 03:34:08.287424 4980 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 03:34:08 crc kubenswrapper[4980]: I1206 03:34:08.287480 4980 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 03:34:08 crc kubenswrapper[4980]: I1206 03:34:08.289124 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:34:08 crc kubenswrapper[4980]: I1206 03:34:08.289150 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:34:08 crc kubenswrapper[4980]: I1206 03:34:08.289172 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:34:08 crc kubenswrapper[4980]: I1206 03:34:08.289184 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:34:08 crc kubenswrapper[4980]: I1206 03:34:08.289190 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:34:08 crc kubenswrapper[4980]: I1206 03:34:08.289209 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:34:09 crc kubenswrapper[4980]: I1206 03:34:09.290988 4980 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 03:34:09 crc kubenswrapper[4980]: I1206 03:34:09.291128 4980 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 03:34:09 crc kubenswrapper[4980]: I1206 03:34:09.292450 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:34:09 crc kubenswrapper[4980]: I1206 03:34:09.292493 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:34:09 crc kubenswrapper[4980]: I1206 03:34:09.292503 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:34:09 crc kubenswrapper[4980]: I1206 03:34:09.292714 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:34:09 crc kubenswrapper[4980]: I1206 03:34:09.292762 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:34:09 crc kubenswrapper[4980]: I1206 03:34:09.292789 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:34:10 crc kubenswrapper[4980]: I1206 03:34:10.255484 4980 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 06 03:34:10 crc kubenswrapper[4980]: I1206 03:34:10.255854 4980 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 03:34:10 crc kubenswrapper[4980]: I1206 03:34:10.257842 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:34:10 crc kubenswrapper[4980]: I1206 03:34:10.257901 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:34:10 crc kubenswrapper[4980]: I1206 03:34:10.257924 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:34:10 crc kubenswrapper[4980]: I1206 03:34:10.395892 4980 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-etcd/etcd-crc" Dec 06 03:34:10 crc kubenswrapper[4980]: I1206 03:34:10.396122 4980 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 03:34:10 crc kubenswrapper[4980]: I1206 03:34:10.397801 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:34:10 crc kubenswrapper[4980]: I1206 03:34:10.397852 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:34:10 crc kubenswrapper[4980]: I1206 03:34:10.397870 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:34:10 crc kubenswrapper[4980]: I1206 03:34:10.540876 4980 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 06 03:34:10 crc kubenswrapper[4980]: I1206 03:34:10.541114 4980 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 03:34:10 crc kubenswrapper[4980]: I1206 03:34:10.542838 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:34:10 crc kubenswrapper[4980]: I1206 03:34:10.542888 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:34:10 crc kubenswrapper[4980]: I1206 03:34:10.542904 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:34:10 crc kubenswrapper[4980]: I1206 03:34:10.549085 4980 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 06 03:34:11 crc kubenswrapper[4980]: I1206 03:34:11.252201 4980 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 06 03:34:11 crc kubenswrapper[4980]: I1206 03:34:11.252341 4980 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 06 03:34:11 crc kubenswrapper[4980]: E1206 03:34:11.261623 4980 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Dec 06 03:34:11 crc kubenswrapper[4980]: I1206 03:34:11.296679 4980 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 03:34:11 crc kubenswrapper[4980]: I1206 03:34:11.298112 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:34:11 crc kubenswrapper[4980]: I1206 03:34:11.298200 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:34:11 crc kubenswrapper[4980]: I1206 03:34:11.298226 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:34:13 crc kubenswrapper[4980]: I1206 03:34:13.607731 4980 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 06 03:34:13 crc kubenswrapper[4980]: I1206 03:34:13.607942 4980 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 03:34:13 crc kubenswrapper[4980]: I1206 03:34:13.609945 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:34:13 crc kubenswrapper[4980]: I1206 03:34:13.610023 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:34:13 crc kubenswrapper[4980]: I1206 03:34:13.610050 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:34:15 crc kubenswrapper[4980]: I1206 03:34:15.105798 4980 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": net/http: TLS handshake timeout Dec 06 03:34:15 crc kubenswrapper[4980]: W1206 03:34:15.112625 4980 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": net/http: TLS handshake timeout Dec 06 03:34:15 crc kubenswrapper[4980]: I1206 03:34:15.112817 4980 trace.go:236] Trace[602485691]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (06-Dec-2025 03:34:05.110) (total time: 10002ms): Dec 06 03:34:15 crc kubenswrapper[4980]: Trace[602485691]: ---"Objects listed" error:Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": net/http: TLS handshake timeout 10001ms (03:34:15.112) Dec 06 03:34:15 crc kubenswrapper[4980]: Trace[602485691]: [10.002032032s] [10.002032032s] END Dec 06 03:34:15 crc kubenswrapper[4980]: E1206 03:34:15.112858 4980 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": net/http: TLS handshake timeout" logger="UnhandledError" Dec 06 03:34:16 crc kubenswrapper[4980]: I1206 03:34:16.621246 4980 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Dec 06 03:34:16 crc kubenswrapper[4980]: I1206 03:34:16.621376 4980 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Dec 06 03:34:16 crc kubenswrapper[4980]: I1206 03:34:16.632555 4980 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Dec 06 03:34:16 crc kubenswrapper[4980]: I1206 03:34:16.632653 4980 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Dec 06 03:34:18 crc kubenswrapper[4980]: I1206 03:34:18.114201 4980 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[+]ping ok Dec 06 03:34:18 crc kubenswrapper[4980]: [+]log ok Dec 06 03:34:18 crc kubenswrapper[4980]: [+]etcd ok Dec 06 03:34:18 crc kubenswrapper[4980]: [+]poststarthook/openshift.io-oauth-apiserver-reachable ok Dec 06 03:34:18 crc kubenswrapper[4980]: [+]poststarthook/start-apiserver-admission-initializer ok Dec 06 03:34:18 crc kubenswrapper[4980]: [+]poststarthook/quota.openshift.io-clusterquotamapping ok Dec 06 03:34:18 crc kubenswrapper[4980]: [+]poststarthook/openshift.io-api-request-count-filter ok Dec 06 03:34:18 crc kubenswrapper[4980]: [+]poststarthook/openshift.io-startkubeinformers ok Dec 06 03:34:18 crc kubenswrapper[4980]: [+]poststarthook/openshift.io-openshift-apiserver-reachable ok Dec 06 03:34:18 crc kubenswrapper[4980]: [+]poststarthook/generic-apiserver-start-informers ok Dec 06 03:34:18 crc kubenswrapper[4980]: [+]poststarthook/priority-and-fairness-config-consumer ok Dec 06 03:34:18 crc kubenswrapper[4980]: [+]poststarthook/priority-and-fairness-filter ok Dec 06 03:34:18 crc kubenswrapper[4980]: [+]poststarthook/storage-object-count-tracker-hook ok Dec 06 03:34:18 crc kubenswrapper[4980]: [+]poststarthook/start-apiextensions-informers ok Dec 06 03:34:18 crc kubenswrapper[4980]: [+]poststarthook/start-apiextensions-controllers ok Dec 06 03:34:18 crc kubenswrapper[4980]: [+]poststarthook/crd-informer-synced ok Dec 06 03:34:18 crc kubenswrapper[4980]: [+]poststarthook/start-system-namespaces-controller ok Dec 06 03:34:18 crc kubenswrapper[4980]: [+]poststarthook/start-cluster-authentication-info-controller ok Dec 06 03:34:18 crc kubenswrapper[4980]: [+]poststarthook/start-kube-apiserver-identity-lease-controller ok Dec 06 03:34:18 crc kubenswrapper[4980]: [+]poststarthook/start-kube-apiserver-identity-lease-garbage-collector ok Dec 06 03:34:18 crc kubenswrapper[4980]: [+]poststarthook/start-legacy-token-tracking-controller ok Dec 06 03:34:18 crc kubenswrapper[4980]: [+]poststarthook/start-service-ip-repair-controllers ok Dec 06 03:34:18 crc kubenswrapper[4980]: [-]poststarthook/rbac/bootstrap-roles failed: reason withheld Dec 06 03:34:18 crc kubenswrapper[4980]: [+]poststarthook/scheduling/bootstrap-system-priority-classes ok Dec 06 03:34:18 crc kubenswrapper[4980]: [+]poststarthook/priority-and-fairness-config-producer ok Dec 06 03:34:18 crc kubenswrapper[4980]: [+]poststarthook/bootstrap-controller ok Dec 06 03:34:18 crc kubenswrapper[4980]: [+]poststarthook/aggregator-reload-proxy-client-cert ok Dec 06 03:34:18 crc kubenswrapper[4980]: [+]poststarthook/start-kube-aggregator-informers ok Dec 06 03:34:18 crc kubenswrapper[4980]: [+]poststarthook/apiservice-status-local-available-controller ok Dec 06 03:34:18 crc kubenswrapper[4980]: [+]poststarthook/apiservice-status-remote-available-controller ok Dec 06 03:34:18 crc kubenswrapper[4980]: [+]poststarthook/apiservice-registration-controller ok Dec 06 03:34:18 crc kubenswrapper[4980]: [+]poststarthook/apiservice-wait-for-first-sync ok Dec 06 03:34:18 crc kubenswrapper[4980]: [+]poststarthook/apiservice-discovery-controller ok Dec 06 03:34:18 crc kubenswrapper[4980]: [+]poststarthook/kube-apiserver-autoregistration ok Dec 06 03:34:18 crc kubenswrapper[4980]: [+]autoregister-completion ok Dec 06 03:34:18 crc kubenswrapper[4980]: [+]poststarthook/apiservice-openapi-controller ok Dec 06 03:34:18 crc kubenswrapper[4980]: [+]poststarthook/apiservice-openapiv3-controller ok Dec 06 03:34:18 crc kubenswrapper[4980]: livez check failed Dec 06 03:34:18 crc kubenswrapper[4980]: I1206 03:34:18.114341 4980 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 06 03:34:20 crc kubenswrapper[4980]: I1206 03:34:20.424592 4980 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-etcd/etcd-crc" Dec 06 03:34:20 crc kubenswrapper[4980]: I1206 03:34:20.425044 4980 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 03:34:20 crc kubenswrapper[4980]: I1206 03:34:20.426808 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:34:20 crc kubenswrapper[4980]: I1206 03:34:20.426904 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:34:20 crc kubenswrapper[4980]: I1206 03:34:20.426920 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:34:20 crc kubenswrapper[4980]: I1206 03:34:20.442433 4980 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-etcd/etcd-crc" Dec 06 03:34:21 crc kubenswrapper[4980]: I1206 03:34:21.252670 4980 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 06 03:34:21 crc kubenswrapper[4980]: I1206 03:34:21.252724 4980 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 06 03:34:21 crc kubenswrapper[4980]: E1206 03:34:21.262038 4980 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Dec 06 03:34:21 crc kubenswrapper[4980]: I1206 03:34:21.320974 4980 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Dec 06 03:34:21 crc kubenswrapper[4980]: E1206 03:34:21.619452 4980 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": context deadline exceeded" interval="6.4s" Dec 06 03:34:21 crc kubenswrapper[4980]: I1206 03:34:21.632882 4980 trace.go:236] Trace[214286202]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (06-Dec-2025 03:34:08.815) (total time: 12817ms): Dec 06 03:34:21 crc kubenswrapper[4980]: Trace[214286202]: ---"Objects listed" error: 12817ms (03:34:21.632) Dec 06 03:34:21 crc kubenswrapper[4980]: Trace[214286202]: [12.817112396s] [12.817112396s] END Dec 06 03:34:21 crc kubenswrapper[4980]: I1206 03:34:21.632912 4980 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Dec 06 03:34:21 crc kubenswrapper[4980]: I1206 03:34:21.633049 4980 trace.go:236] Trace[513453423]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (06-Dec-2025 03:34:07.901) (total time: 13731ms): Dec 06 03:34:21 crc kubenswrapper[4980]: Trace[513453423]: ---"Objects listed" error: 13731ms (03:34:21.633) Dec 06 03:34:21 crc kubenswrapper[4980]: Trace[513453423]: [13.731601033s] [13.731601033s] END Dec 06 03:34:21 crc kubenswrapper[4980]: I1206 03:34:21.633058 4980 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Dec 06 03:34:21 crc kubenswrapper[4980]: I1206 03:34:21.657791 4980 trace.go:236] Trace[1806786169]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (06-Dec-2025 03:34:10.345) (total time: 11311ms): Dec 06 03:34:21 crc kubenswrapper[4980]: Trace[1806786169]: ---"Objects listed" error: 11311ms (03:34:21.657) Dec 06 03:34:21 crc kubenswrapper[4980]: Trace[1806786169]: [11.311745615s] [11.311745615s] END Dec 06 03:34:21 crc kubenswrapper[4980]: I1206 03:34:21.657843 4980 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Dec 06 03:34:21 crc kubenswrapper[4980]: I1206 03:34:21.732275 4980 reconstruct.go:205] "DevicePaths of reconstructed volumes updated" Dec 06 03:34:21 crc kubenswrapper[4980]: E1206 03:34:21.808554 4980 kubelet_node_status.go:99] "Unable to register node with API server" err="nodes \"crc\" is forbidden: autoscaling.openshift.io/ManagedNode infra config cache not synchronized" node="crc" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.053015 4980 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:36896->192.168.126.11:17697: read: connection reset by peer" start-of-body= Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.053477 4980 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:36896->192.168.126.11:17697: read: connection reset by peer" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.113925 4980 apiserver.go:52] "Watching apiserver" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.117757 4980 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.121924 4980 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd/etcd-crc","openshift-network-console/networking-console-plugin-85b44fc459-gdk6g","openshift-network-diagnostics/network-check-source-55646444c4-trplf","openshift-network-diagnostics/network-check-target-xd92c","openshift-network-node-identity/network-node-identity-vrzqb","openshift-network-operator/iptables-alerter-4ln5h","openshift-network-operator/network-operator-58b4c7f79c-55gtf"] Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.122658 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.122714 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 03:34:22 crc kubenswrapper[4980]: E1206 03:34:22.122737 4980 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 03:34:22 crc kubenswrapper[4980]: E1206 03:34:22.122832 4980 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.123112 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.123437 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.124117 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.124781 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 03:34:22 crc kubenswrapper[4980]: E1206 03:34:22.125174 4980 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.130397 4980 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.130613 4980 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.130768 4980 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.130788 4980 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.130892 4980 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.139317 4980 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.145338 4980 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.149443 4980 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.149735 4980 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.215449 4980 desired_state_of_world_populator.go:154] "Finished populating initial desired state of world" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.236415 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.236466 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.236495 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.236547 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.236569 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.236598 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.236622 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.236642 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.236661 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.236681 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.236702 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.236723 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.236742 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.236761 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.236798 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.236820 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.236842 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") pod \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\" (UID: \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\") " Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.236866 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.236886 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.236924 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.236961 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.236982 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.237002 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.237038 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.237076 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.237096 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") pod \"44663579-783b-4372-86d6-acf235a62d72\" (UID: \"44663579-783b-4372-86d6-acf235a62d72\") " Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.237141 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.237172 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.237203 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.237250 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.237280 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.237311 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.237332 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.237359 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.237367 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.237382 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.237458 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.237486 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.237502 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.237535 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.237561 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.237576 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.237590 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.237634 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.237651 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.237669 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.237683 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.237700 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.237725 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.237741 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.237756 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.237776 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.237792 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.237809 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.237811 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.237824 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.237840 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.237856 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.237870 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.237886 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.237900 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.237915 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.237934 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.237948 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.237962 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.237976 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.237991 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.238007 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.238027 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.238042 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.238040 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.238060 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.238080 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.238096 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.238128 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.238143 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.238159 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.238175 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.238191 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.238207 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.238226 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.238240 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.238247 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.238256 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.238304 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.238329 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.238349 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.238371 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.238404 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.238430 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.238451 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.238473 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") pod \"49ef4625-1d3a-4a9f-b595-c2433d32326d\" (UID: \"49ef4625-1d3a-4a9f-b595-c2433d32326d\") " Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.238497 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.238547 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.238578 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.238600 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.238626 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.238648 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.238649 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" (OuterVolumeSpecName: "client-ca") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.238669 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.238694 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.238716 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.238739 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.238761 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.238782 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.238798 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.238804 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.238833 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.238851 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.238867 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.238887 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.238904 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.238924 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.238939 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.238957 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.238972 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.238986 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.239001 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.239002 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" (OuterVolumeSpecName: "kube-api-access-d4lsv") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "kube-api-access-d4lsv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.239017 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.239064 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.239087 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.239135 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.239158 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.239179 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.239203 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.239224 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.239246 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.239270 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.239294 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.239316 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.239329 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" (OuterVolumeSpecName: "service-ca") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.239338 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.239369 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.239398 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.239421 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.239445 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.239469 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.239493 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.239562 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.239586 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.239606 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.239609 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.239640 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.239657 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.239679 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.239697 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.239714 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.239731 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.239746 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.239768 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.239785 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.239808 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.239825 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.239851 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.239870 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.239885 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.239925 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.239944 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.239962 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.239979 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.239998 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.240015 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.240031 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.240046 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.240062 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.240083 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.240097 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.240113 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") pod \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\" (UID: \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\") " Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.240128 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.240144 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.240160 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.240176 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.240192 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.240209 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.240225 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.240241 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.240258 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.240274 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.240291 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.240318 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.240316 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.240343 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.240361 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.240377 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.240394 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.240411 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.240429 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.240444 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.240461 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.240478 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.240494 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.240720 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" (OuterVolumeSpecName: "kube-api-access-v47cf") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "kube-api-access-v47cf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.240916 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" (OuterVolumeSpecName: "kube-api-access-2w9zh") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "kube-api-access-2w9zh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.241171 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" (OuterVolumeSpecName: "default-certificate") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "default-certificate". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.241229 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" (OuterVolumeSpecName: "config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.241376 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" (OuterVolumeSpecName: "kube-api-access-pjr6v") pod "49ef4625-1d3a-4a9f-b595-c2433d32326d" (UID: "49ef4625-1d3a-4a9f-b595-c2433d32326d"). InnerVolumeSpecName "kube-api-access-pjr6v". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.241436 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.241481 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" (OuterVolumeSpecName: "package-server-manager-serving-cert") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "package-server-manager-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.242273 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" (OuterVolumeSpecName: "kube-api-access-pj782") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "kube-api-access-pj782". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.242290 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" (OuterVolumeSpecName: "kube-api-access-lzf88") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "kube-api-access-lzf88". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.241584 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.241671 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" (OuterVolumeSpecName: "kube-api-access-w9rds") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "kube-api-access-w9rds". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.241803 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" (OuterVolumeSpecName: "kube-api-access-x7zkh") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "kube-api-access-x7zkh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.241853 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" (OuterVolumeSpecName: "kube-api-access-vt5rc") pod "44663579-783b-4372-86d6-acf235a62d72" (UID: "44663579-783b-4372-86d6-acf235a62d72"). InnerVolumeSpecName "kube-api-access-vt5rc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.241910 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.242005 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.242169 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" (OuterVolumeSpecName: "kube-api-access-xcgwh") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "kube-api-access-xcgwh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.242310 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" (OuterVolumeSpecName: "config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.242531 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" (OuterVolumeSpecName: "stats-auth") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "stats-auth". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.242541 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" (OuterVolumeSpecName: "kube-api-access-kfwg7") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "kube-api-access-kfwg7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.242590 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" (OuterVolumeSpecName: "kube-api-access-249nr") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "kube-api-access-249nr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.242617 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.242699 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" (OuterVolumeSpecName: "utilities") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.242762 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.242784 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.242882 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" (OuterVolumeSpecName: "utilities") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.242929 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" (OuterVolumeSpecName: "mcc-auth-proxy-config") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "mcc-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.242956 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" (OuterVolumeSpecName: "kube-api-access-w7l8j") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "kube-api-access-w7l8j". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.242976 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.243068 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" (OuterVolumeSpecName: "kube-api-access-jhbk2") pod "bd23aa5c-e532-4e53-bccf-e79f130c5ae8" (UID: "bd23aa5c-e532-4e53-bccf-e79f130c5ae8"). InnerVolumeSpecName "kube-api-access-jhbk2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.243101 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.243316 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.243347 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" (OuterVolumeSpecName: "kube-api-access-tk88c") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "kube-api-access-tk88c". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.243377 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" (OuterVolumeSpecName: "kube-api-access-lz9wn") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "kube-api-access-lz9wn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.243386 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.243453 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.243482 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" (OuterVolumeSpecName: "signing-cabundle") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-cabundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.243487 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" (OuterVolumeSpecName: "kube-api-access-x4zgh") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "kube-api-access-x4zgh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.243762 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" (OuterVolumeSpecName: "kube-api-access-ngvvp") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "kube-api-access-ngvvp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.243817 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" (OuterVolumeSpecName: "cni-sysctl-allowlist") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-sysctl-allowlist". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.243868 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.243959 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" (OuterVolumeSpecName: "kube-api-access-6g6sz") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "kube-api-access-6g6sz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.243976 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" (OuterVolumeSpecName: "kube-api-access-htfz6") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "kube-api-access-htfz6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.244130 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" (OuterVolumeSpecName: "image-import-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "image-import-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.244144 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" (OuterVolumeSpecName: "kube-api-access-sb6h7") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "kube-api-access-sb6h7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.243998 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.244226 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" (OuterVolumeSpecName: "config") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.244324 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.244352 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" (OuterVolumeSpecName: "config") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.244384 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.244395 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" (OuterVolumeSpecName: "kube-api-access-fqsjt") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "kube-api-access-fqsjt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.244494 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.244525 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" (OuterVolumeSpecName: "control-plane-machine-set-operator-tls") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "control-plane-machine-set-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.244584 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" (OuterVolumeSpecName: "certs") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.244651 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" (OuterVolumeSpecName: "config") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.244727 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.241502 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.244749 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" (OuterVolumeSpecName: "kube-api-access-zkvpv") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "kube-api-access-zkvpv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.244796 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.244825 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.244850 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.244873 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.244898 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.244923 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.244946 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.244971 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.245007 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.245030 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.245055 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.245088 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.245120 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.245183 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.245216 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.245245 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.245270 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.245298 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.245322 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.245353 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.245377 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.245404 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.245444 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.245480 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.245509 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.246422 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.246497 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.246828 4980 reconciler_common.go:293] "Volume detached for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") on node \"crc\" DevicePath \"\"" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.246845 4980 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") on node \"crc\" DevicePath \"\"" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.246863 4980 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") on node \"crc\" DevicePath \"\"" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.246882 4980 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") on node \"crc\" DevicePath \"\"" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.246900 4980 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") on node \"crc\" DevicePath \"\"" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.246919 4980 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.246941 4980 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.246959 4980 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") on node \"crc\" DevicePath \"\"" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.246976 4980 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") on node \"crc\" DevicePath \"\"" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.246994 4980 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") on node \"crc\" DevicePath \"\"" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.247011 4980 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") on node \"crc\" DevicePath \"\"" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.247028 4980 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") on node \"crc\" DevicePath \"\"" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.247046 4980 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") on node \"crc\" DevicePath \"\"" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.247064 4980 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.247081 4980 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") on node \"crc\" DevicePath \"\"" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.247099 4980 reconciler_common.go:293] "Volume detached for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") on node \"crc\" DevicePath \"\"" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.247115 4980 reconciler_common.go:293] "Volume detached for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") on node \"crc\" DevicePath \"\"" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.247133 4980 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") on node \"crc\" DevicePath \"\"" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.247150 4980 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") on node \"crc\" DevicePath \"\"" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.247190 4980 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") on node \"crc\" DevicePath \"\"" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.247212 4980 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") on node \"crc\" DevicePath \"\"" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.247230 4980 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") on node \"crc\" DevicePath \"\"" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.247248 4980 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") on node \"crc\" DevicePath \"\"" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.247266 4980 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") on node \"crc\" DevicePath \"\"" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.247287 4980 reconciler_common.go:293] "Volume detached for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.247305 4980 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") on node \"crc\" DevicePath \"\"" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.247324 4980 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.247344 4980 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.247362 4980 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.247379 4980 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.247395 4980 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") on node \"crc\" DevicePath \"\"" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.247417 4980 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") on node \"crc\" DevicePath \"\"" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.247436 4980 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.247454 4980 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.247472 4980 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") on node \"crc\" DevicePath \"\"" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.247489 4980 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") on node \"crc\" DevicePath \"\"" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.247510 4980 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.247550 4980 reconciler_common.go:293] "Volume detached for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") on node \"crc\" DevicePath \"\"" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.247567 4980 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.247609 4980 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.247628 4980 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.247647 4980 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.247664 4980 reconciler_common.go:293] "Volume detached for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") on node \"crc\" DevicePath \"\"" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.247681 4980 reconciler_common.go:293] "Volume detached for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") on node \"crc\" DevicePath \"\"" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.247712 4980 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") on node \"crc\" DevicePath \"\"" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.247729 4980 reconciler_common.go:293] "Volume detached for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.247753 4980 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") on node \"crc\" DevicePath \"\"" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.247772 4980 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") on node \"crc\" DevicePath \"\"" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.247790 4980 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") on node \"crc\" DevicePath \"\"" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.247807 4980 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") on node \"crc\" DevicePath \"\"" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.247825 4980 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") on node \"crc\" DevicePath \"\"" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.247842 4980 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.247859 4980 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") on node \"crc\" DevicePath \"\"" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.247877 4980 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") on node \"crc\" DevicePath \"\"" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.247895 4980 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") on node \"crc\" DevicePath \"\"" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.247911 4980 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") on node \"crc\" DevicePath \"\"" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.247934 4980 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") on node \"crc\" DevicePath \"\"" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.247953 4980 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") on node \"crc\" DevicePath \"\"" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.247970 4980 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") on node \"crc\" DevicePath \"\"" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.247989 4980 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") on node \"crc\" DevicePath \"\"" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.248006 4980 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") on node \"crc\" DevicePath \"\"" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.248024 4980 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.248042 4980 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") on node \"crc\" DevicePath \"\"" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.248061 4980 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") on node \"crc\" DevicePath \"\"" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.248078 4980 reconciler_common.go:293] "Volume detached for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") on node \"crc\" DevicePath \"\"" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.248096 4980 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") on node \"crc\" DevicePath \"\"" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.248115 4980 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") on node \"crc\" DevicePath \"\"" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.248772 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.249648 4980 swap_util.go:74] "error creating dir to test if tmpfs noswap is enabled. Assuming not supported" mount path="" error="stat /var/lib/kubelet/plugins/kubernetes.io/empty-dir: no such file or directory" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.244973 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" (OuterVolumeSpecName: "etcd-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.252574 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" (OuterVolumeSpecName: "kube-api-access-7c4vf") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "kube-api-access-7c4vf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.245191 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.245361 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.245514 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.245511 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" (OuterVolumeSpecName: "kube-api-access-pcxfs") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "kube-api-access-pcxfs". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.245129 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" (OuterVolumeSpecName: "config-volume") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 03:34:22 crc kubenswrapper[4980]: E1206 03:34:22.245697 4980 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.245749 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" (OuterVolumeSpecName: "signing-key") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.245761 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" (OuterVolumeSpecName: "audit") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "audit". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.245909 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" (OuterVolumeSpecName: "kube-api-access-x2m85") pod "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" (UID: "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d"). InnerVolumeSpecName "kube-api-access-x2m85". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.245936 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.246057 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.246110 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.246174 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.246545 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" (OuterVolumeSpecName: "kube-api-access-gf66m") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "kube-api-access-gf66m". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.246615 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" (OuterVolumeSpecName: "mcd-auth-proxy-config") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "mcd-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.247321 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" (OuterVolumeSpecName: "kube-api-access-wxkg8") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "kube-api-access-wxkg8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.247390 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" (OuterVolumeSpecName: "node-bootstrap-token") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "node-bootstrap-token". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.247650 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" (OuterVolumeSpecName: "multus-daemon-config") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "multus-daemon-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.247905 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" (OuterVolumeSpecName: "config") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.247765 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.248315 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.248470 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.248761 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.248877 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" (OuterVolumeSpecName: "etcd-service-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.249066 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.249082 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.249432 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.249455 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.249795 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" (OuterVolumeSpecName: "kube-api-access-s4n52") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "kube-api-access-s4n52". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.249852 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" (OuterVolumeSpecName: "kube-api-access-dbsvg") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "kube-api-access-dbsvg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.250020 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" (OuterVolumeSpecName: "machine-approver-tls") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "machine-approver-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.250057 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.250185 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" (OuterVolumeSpecName: "kube-api-access-mg5zb") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "kube-api-access-mg5zb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.250301 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" (OuterVolumeSpecName: "kube-api-access-zgdk5") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "kube-api-access-zgdk5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.250667 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.250673 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" (OuterVolumeSpecName: "kube-api-access-d6qdx") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "kube-api-access-d6qdx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.250724 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.250882 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" (OuterVolumeSpecName: "kube-api-access-fcqwp") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "kube-api-access-fcqwp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.250924 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" (OuterVolumeSpecName: "kube-api-access-nzwt7") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "kube-api-access-nzwt7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.251150 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" (OuterVolumeSpecName: "kube-api-access-6ccd8") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "kube-api-access-6ccd8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.251311 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" (OuterVolumeSpecName: "kube-api-access-qs4fp") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "kube-api-access-qs4fp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.251493 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" (OuterVolumeSpecName: "kube-api-access-2d4wz") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "kube-api-access-2d4wz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.251943 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" (OuterVolumeSpecName: "kube-api-access-279lb") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "kube-api-access-279lb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.252080 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" (OuterVolumeSpecName: "kube-api-access-4d4hj") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "kube-api-access-4d4hj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.252144 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" (OuterVolumeSpecName: "kube-api-access-xcphl") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "kube-api-access-xcphl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.252474 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.252487 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" (OuterVolumeSpecName: "tmpfs") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "tmpfs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.252558 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" (OuterVolumeSpecName: "config") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.252768 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.252886 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.252991 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" (OuterVolumeSpecName: "kube-api-access-jkwtn") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "kube-api-access-jkwtn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.253081 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.253201 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.253277 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" (OuterVolumeSpecName: "utilities") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.253340 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" (OuterVolumeSpecName: "service-ca") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.253439 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.253538 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.253564 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.253567 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" (OuterVolumeSpecName: "console-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.253765 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" (OuterVolumeSpecName: "cert") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.253775 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.253850 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.253986 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" (OuterVolumeSpecName: "kube-api-access-bf2bz") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "kube-api-access-bf2bz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.253993 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.254063 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" (OuterVolumeSpecName: "webhook-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "webhook-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.254081 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.254272 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.254369 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" (OuterVolumeSpecName: "kube-api-access-cfbct") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "kube-api-access-cfbct". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.254486 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.254625 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" (OuterVolumeSpecName: "image-registry-operator-tls") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "image-registry-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.254773 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" (OuterVolumeSpecName: "config") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.254779 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.255084 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" (OuterVolumeSpecName: "available-featuregates") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "available-featuregates". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.255309 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.255734 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" (OuterVolumeSpecName: "images") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.255811 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.255940 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" (OuterVolumeSpecName: "machine-api-operator-tls") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "machine-api-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.256594 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" (OuterVolumeSpecName: "config") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.257038 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.257161 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" (OuterVolumeSpecName: "config") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.257404 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 06 03:34:22 crc kubenswrapper[4980]: E1206 03:34:22.257486 4980 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.257490 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.257585 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" (OuterVolumeSpecName: "serviceca") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "serviceca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 03:34:22 crc kubenswrapper[4980]: E1206 03:34:22.257761 4980 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-06 03:34:22.757741873 +0000 UTC m=+22.002493144 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.257880 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" (OuterVolumeSpecName: "client-ca") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 03:34:22 crc kubenswrapper[4980]: E1206 03:34:22.258126 4980 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-06 03:34:22.757958439 +0000 UTC m=+22.002709800 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.258489 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.258635 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.259503 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.259819 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" (OuterVolumeSpecName: "config") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.259896 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" (OuterVolumeSpecName: "kube-api-access-qg5z5") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "kube-api-access-qg5z5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.260275 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" (OuterVolumeSpecName: "config") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.259811 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.260774 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.261136 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.261494 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.261635 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" (OuterVolumeSpecName: "config") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.261870 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.262097 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.262456 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" (OuterVolumeSpecName: "webhook-certs") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "webhook-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.262505 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.262772 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.262859 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" (OuterVolumeSpecName: "samples-operator-tls") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "samples-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.263070 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" (OuterVolumeSpecName: "kube-api-access-9xfj7") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "kube-api-access-9xfj7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.263140 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" (OuterVolumeSpecName: "kube-api-access-rnphk") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "kube-api-access-rnphk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.263412 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.263426 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" (OuterVolumeSpecName: "kube-api-access-w4xd4") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "kube-api-access-w4xd4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.257910 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.263724 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.263942 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" (OuterVolumeSpecName: "ovn-control-plane-metrics-cert") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovn-control-plane-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.263999 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.272028 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.273187 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.274442 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" (OuterVolumeSpecName: "kube-api-access-mnrrd") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "kube-api-access-mnrrd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.274631 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.274972 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.275201 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.276960 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.276962 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" (OuterVolumeSpecName: "apiservice-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "apiservice-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.277576 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" (OuterVolumeSpecName: "images") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.277628 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.277965 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.278218 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" (OuterVolumeSpecName: "kube-api-access-8tdtz") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "kube-api-access-8tdtz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.278455 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" (OuterVolumeSpecName: "utilities") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.257905 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" (OuterVolumeSpecName: "config") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.278530 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 03:34:22 crc kubenswrapper[4980]: E1206 03:34:22.283797 4980 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 03:34:22.783767562 +0000 UTC m=+22.028518833 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.285741 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.292315 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" (OuterVolumeSpecName: "config") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 03:34:22 crc kubenswrapper[4980]: E1206 03:34:22.353998 4980 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 06 03:34:22 crc kubenswrapper[4980]: E1206 03:34:22.354028 4980 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 06 03:34:22 crc kubenswrapper[4980]: E1206 03:34:22.354038 4980 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 06 03:34:22 crc kubenswrapper[4980]: E1206 03:34:22.354093 4980 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-06 03:34:22.854077108 +0000 UTC m=+22.098828379 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 06 03:34:22 crc kubenswrapper[4980]: E1206 03:34:22.354862 4980 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 06 03:34:22 crc kubenswrapper[4980]: E1206 03:34:22.354908 4980 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 06 03:34:22 crc kubenswrapper[4980]: E1206 03:34:22.354924 4980 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.355029 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.355053 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.355153 4980 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") on node \"crc\" DevicePath \"\"" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.355168 4980 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") on node \"crc\" DevicePath \"\"" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.355185 4980 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") on node \"crc\" DevicePath \"\"" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.355195 4980 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.355116 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 06 03:34:22 crc kubenswrapper[4980]: E1206 03:34:22.355245 4980 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-06 03:34:22.85523542 +0000 UTC m=+22.099986691 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.355285 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.355208 4980 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") on node \"crc\" DevicePath \"\"" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.355329 4980 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") on node \"crc\" DevicePath \"\"" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.355338 4980 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.355348 4980 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.355358 4980 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") on node \"crc\" DevicePath \"\"" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.355367 4980 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") on node \"crc\" DevicePath \"\"" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.355375 4980 reconciler_common.go:293] "Volume detached for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") on node \"crc\" DevicePath \"\"" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.355385 4980 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.355394 4980 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.355403 4980 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") on node \"crc\" DevicePath \"\"" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.355419 4980 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.355428 4980 reconciler_common.go:293] "Volume detached for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") on node \"crc\" DevicePath \"\"" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.355437 4980 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.355446 4980 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.355454 4980 reconciler_common.go:293] "Volume detached for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") on node \"crc\" DevicePath \"\"" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.355469 4980 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") on node \"crc\" DevicePath \"\"" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.355479 4980 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.355488 4980 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") on node \"crc\" DevicePath \"\"" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.355495 4980 reconciler_common.go:293] "Volume detached for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") on node \"crc\" DevicePath \"\"" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.355509 4980 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.355562 4980 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.355571 4980 reconciler_common.go:293] "Volume detached for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") on node \"crc\" DevicePath \"\"" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.355580 4980 reconciler_common.go:293] "Volume detached for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.355588 4980 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") on node \"crc\" DevicePath \"\"" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.355596 4980 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") on node \"crc\" DevicePath \"\"" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.355609 4980 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.355617 4980 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.355630 4980 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") on node \"crc\" DevicePath \"\"" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.355640 4980 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.355653 4980 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.355667 4980 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") on node \"crc\" DevicePath \"\"" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.355682 4980 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") on node \"crc\" DevicePath \"\"" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.355691 4980 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") on node \"crc\" DevicePath \"\"" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.355699 4980 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") on node \"crc\" DevicePath \"\"" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.355708 4980 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") on node \"crc\" DevicePath \"\"" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.355717 4980 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") on node \"crc\" DevicePath \"\"" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.355725 4980 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") on node \"crc\" DevicePath \"\"" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.355733 4980 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") on node \"crc\" DevicePath \"\"" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.355743 4980 reconciler_common.go:293] "Volume detached for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") on node \"crc\" DevicePath \"\"" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.355751 4980 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") on node \"crc\" DevicePath \"\"" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.355758 4980 reconciler_common.go:293] "Volume detached for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") on node \"crc\" DevicePath \"\"" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.355766 4980 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") on node \"crc\" DevicePath \"\"" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.355774 4980 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.355783 4980 reconciler_common.go:293] "Volume detached for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") on node \"crc\" DevicePath \"\"" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.355791 4980 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") on node \"crc\" DevicePath \"\"" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.355798 4980 reconciler_common.go:293] "Volume detached for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") on node \"crc\" DevicePath \"\"" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.355806 4980 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.355814 4980 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.355827 4980 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") on node \"crc\" DevicePath \"\"" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.355835 4980 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.355848 4980 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") on node \"crc\" DevicePath \"\"" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.355855 4980 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") on node \"crc\" DevicePath \"\"" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.355864 4980 reconciler_common.go:293] "Volume detached for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") on node \"crc\" DevicePath \"\"" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.355872 4980 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") on node \"crc\" DevicePath \"\"" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.355880 4980 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") on node \"crc\" DevicePath \"\"" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.355888 4980 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.355896 4980 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.355906 4980 reconciler_common.go:293] "Volume detached for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") on node \"crc\" DevicePath \"\"" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.355921 4980 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.355929 4980 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.355937 4980 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") on node \"crc\" DevicePath \"\"" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.355946 4980 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.355954 4980 reconciler_common.go:293] "Volume detached for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") on node \"crc\" DevicePath \"\"" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.355962 4980 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.355970 4980 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.355978 4980 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.355986 4980 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") on node \"crc\" DevicePath \"\"" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.355996 4980 reconciler_common.go:293] "Volume detached for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") on node \"crc\" DevicePath \"\"" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.356004 4980 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") on node \"crc\" DevicePath \"\"" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.356012 4980 reconciler_common.go:293] "Volume detached for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") on node \"crc\" DevicePath \"\"" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.356020 4980 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") on node \"crc\" DevicePath \"\"" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.356028 4980 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.356037 4980 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") on node \"crc\" DevicePath \"\"" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.356045 4980 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") on node \"crc\" DevicePath \"\"" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.356053 4980 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") on node \"crc\" DevicePath \"\"" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.356061 4980 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") on node \"crc\" DevicePath \"\"" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.356069 4980 reconciler_common.go:293] "Volume detached for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") on node \"crc\" DevicePath \"\"" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.356077 4980 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") on node \"crc\" DevicePath \"\"" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.356087 4980 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") on node \"crc\" DevicePath \"\"" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.356095 4980 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.356104 4980 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") on node \"crc\" DevicePath \"\"" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.356112 4980 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") on node \"crc\" DevicePath \"\"" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.356120 4980 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") on node \"crc\" DevicePath \"\"" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.356160 4980 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.356169 4980 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") on node \"crc\" DevicePath \"\"" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.356176 4980 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.356184 4980 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") on node \"crc\" DevicePath \"\"" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.356192 4980 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.356200 4980 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") on node \"crc\" DevicePath \"\"" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.356208 4980 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") on node \"crc\" DevicePath \"\"" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.356216 4980 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") on node \"crc\" DevicePath \"\"" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.356224 4980 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") on node \"crc\" DevicePath \"\"" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.356232 4980 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.356240 4980 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.356248 4980 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") on node \"crc\" DevicePath \"\"" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.356256 4980 reconciler_common.go:293] "Volume detached for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") on node \"crc\" DevicePath \"\"" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.356264 4980 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") on node \"crc\" DevicePath \"\"" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.356272 4980 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") on node \"crc\" DevicePath \"\"" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.356280 4980 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") on node \"crc\" DevicePath \"\"" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.356287 4980 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") on node \"crc\" DevicePath \"\"" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.356295 4980 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.356302 4980 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.356310 4980 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") on node \"crc\" DevicePath \"\"" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.356317 4980 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") on node \"crc\" DevicePath \"\"" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.356325 4980 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") on node \"crc\" DevicePath \"\"" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.356333 4980 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.361690 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.364386 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.365133 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.365906 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.369190 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.370341 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.356341 4980 reconciler_common.go:293] "Volume detached for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") on node \"crc\" DevicePath \"\"" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.370425 4980 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") on node \"crc\" DevicePath \"\"" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.370436 4980 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") on node \"crc\" DevicePath \"\"" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.370476 4980 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.370487 4980 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.370550 4980 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") on node \"crc\" DevicePath \"\"" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.370560 4980 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") on node \"crc\" DevicePath \"\"" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.370577 4980 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.370586 4980 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.370634 4980 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.370644 4980 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") on node \"crc\" DevicePath \"\"" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.370653 4980 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") on node \"crc\" DevicePath \"\"" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.370663 4980 reconciler_common.go:293] "Volume detached for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") on node \"crc\" DevicePath \"\"" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.370671 4980 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.370857 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.371952 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:22Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.381939 4980 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.397473 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.398568 4980 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="773a0bc2aab73ffc2717e987543defc9b786cc22300aa54b4ba4f3454504e331" exitCode=255 Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.399243 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"773a0bc2aab73ffc2717e987543defc9b786cc22300aa54b4ba4f3454504e331"} Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.400309 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 06 03:34:22 crc kubenswrapper[4980]: W1206 03:34:22.422399 4980 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podef543e1b_8068_4ea3_b32a_61027b32e95d.slice/crio-cb8e711cd9f04199553918c742ba12a89d24f07c2d55bbc2e60e379ef2f83428 WatchSource:0}: Error finding container cb8e711cd9f04199553918c742ba12a89d24f07c2d55bbc2e60e379ef2f83428: Status 404 returned error can't find the container with id cb8e711cd9f04199553918c742ba12a89d24f07c2d55bbc2e60e379ef2f83428 Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.424021 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cab608b3-a02f-475b-9aae-9b8f5d878de9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f7fcc44b42bcbb55509048e3a566bb557333c4a83e33d7b0fba8eb514317909\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c22e7dd8886c2dfbb742900ed86ec7b0c2edf9a1363e74f6666961d928f06c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://473e792b2d0a9644854ee0581c6120fc9c989bc342899ef72b1d056080fd4694\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ee8e936dfc39361d6f99460910591ca8c901a20f0d274bbb6aaebfa8a00ee06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://77bc12768b018625ba11a6bfd66688f926c4b7f3bc1d0e41c6551a18b485ecf3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://27cf07cd96893900da4afecd0d56fd9269bc60c1d23246b4685fcd8756357c19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://27cf07cd96893900da4afecd0d56fd9269bc60c1d23246b4685fcd8756357c19\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c49feb24566a8284378670fa0390710d38c237acc94ffb654f90f3fa4626e4c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c49feb24566a8284378670fa0390710d38c237acc94ffb654f90f3fa4626e4c8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:03Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://bb74d1c3d752a379fa394c450217badb45511fbcafad62f7b618131cab94b29e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bb74d1c3d752a379fa394c450217badb45511fbcafad62f7b618131cab94b29e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:01Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.438689 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:22Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.467820 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:22Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.471997 4980 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.472034 4980 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.472043 4980 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.485923 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:22Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.486924 4980 scope.go:117] "RemoveContainer" containerID="773a0bc2aab73ffc2717e987543defc9b786cc22300aa54b4ba4f3454504e331" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.487165 4980 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.491566 4980 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.496672 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.516814 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.581004 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.587958 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:22Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.612435 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cab608b3-a02f-475b-9aae-9b8f5d878de9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f7fcc44b42bcbb55509048e3a566bb557333c4a83e33d7b0fba8eb514317909\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c22e7dd8886c2dfbb742900ed86ec7b0c2edf9a1363e74f6666961d928f06c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://473e792b2d0a9644854ee0581c6120fc9c989bc342899ef72b1d056080fd4694\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ee8e936dfc39361d6f99460910591ca8c901a20f0d274bbb6aaebfa8a00ee06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://77bc12768b018625ba11a6bfd66688f926c4b7f3bc1d0e41c6551a18b485ecf3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://27cf07cd96893900da4afecd0d56fd9269bc60c1d23246b4685fcd8756357c19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://27cf07cd96893900da4afecd0d56fd9269bc60c1d23246b4685fcd8756357c19\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c49feb24566a8284378670fa0390710d38c237acc94ffb654f90f3fa4626e4c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c49feb24566a8284378670fa0390710d38c237acc94ffb654f90f3fa4626e4c8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:03Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://bb74d1c3d752a379fa394c450217badb45511fbcafad62f7b618131cab94b29e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bb74d1c3d752a379fa394c450217badb45511fbcafad62f7b618131cab94b29e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:01Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.636814 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.639642 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7f958150-549d-4c80-8a7e-23c39e4f9500\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c28b6895b14e73f4f3eb7a8ce93076c05092ba9b58ea0e3e879c65bcfa227a21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://76cff8a2d652df38b04eefc053832baf4cbb5db778ff5e525b18c2825141b118\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://206f6f19fe3197f5d6798cc86c5473bf2decb02a2860ee7ffb5efb5f60952e8c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://773a0bc2aab73ffc2717e987543defc9b786cc22300aa54b4ba4f3454504e331\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://773a0bc2aab73ffc2717e987543defc9b786cc22300aa54b4ba4f3454504e331\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-06T03:34:22Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1206 03:34:15.305330 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1206 03:34:15.307249 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1592846546/tls.crt::/tmp/serving-cert-1592846546/tls.key\\\\\\\"\\\\nI1206 03:34:21.814037 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1206 03:34:21.816756 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1206 03:34:21.816780 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1206 03:34:21.816806 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1206 03:34:21.816813 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1206 03:34:21.831301 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1206 03:34:21.831311 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1206 03:34:21.831324 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 03:34:21.831330 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 03:34:21.831335 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1206 03:34:21.831339 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1206 03:34:21.831342 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1206 03:34:21.831346 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1206 03:34:21.911269 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d40e0e7446f8416bf9da94fc53c01be937f9a51a11eb101157d2ddd4597720d2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:04Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2cba89f2f1bed96d59f183628d900d9b72c91eb472e273c8bed0babcab485910\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2cba89f2f1bed96d59f183628d900d9b72c91eb472e273c8bed0babcab485910\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:01Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.733752 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:22Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.749878 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:22Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.767065 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.778500 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.778577 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 03:34:22 crc kubenswrapper[4980]: E1206 03:34:22.778687 4980 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 06 03:34:22 crc kubenswrapper[4980]: E1206 03:34:22.778741 4980 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-06 03:34:23.778726908 +0000 UTC m=+23.023478179 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 06 03:34:22 crc kubenswrapper[4980]: E1206 03:34:22.778781 4980 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 06 03:34:22 crc kubenswrapper[4980]: E1206 03:34:22.778802 4980 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-06 03:34:23.77879672 +0000 UTC m=+23.023547991 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.951904 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.951991 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 03:34:22 crc kubenswrapper[4980]: I1206 03:34:22.952016 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 03:34:22 crc kubenswrapper[4980]: E1206 03:34:22.952134 4980 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 06 03:34:22 crc kubenswrapper[4980]: E1206 03:34:22.952153 4980 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 06 03:34:22 crc kubenswrapper[4980]: E1206 03:34:22.952167 4980 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 06 03:34:22 crc kubenswrapper[4980]: E1206 03:34:22.952213 4980 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-06 03:34:23.952200424 +0000 UTC m=+23.196951695 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 06 03:34:22 crc kubenswrapper[4980]: E1206 03:34:22.952270 4980 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 03:34:23.952264396 +0000 UTC m=+23.197015667 (durationBeforeRetry 1s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 03:34:22 crc kubenswrapper[4980]: E1206 03:34:22.952310 4980 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 06 03:34:22 crc kubenswrapper[4980]: E1206 03:34:22.952319 4980 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 06 03:34:22 crc kubenswrapper[4980]: E1206 03:34:22.952326 4980 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 06 03:34:22 crc kubenswrapper[4980]: E1206 03:34:22.952355 4980 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-06 03:34:23.952349269 +0000 UTC m=+23.197100540 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 06 03:34:23 crc kubenswrapper[4980]: I1206 03:34:23.007593 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 06 03:34:23 crc kubenswrapper[4980]: I1206 03:34:23.133242 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:22Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 06 03:34:23 crc kubenswrapper[4980]: I1206 03:34:23.135840 4980 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 06 03:34:23 crc kubenswrapper[4980]: I1206 03:34:23.235803 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 03:34:23 crc kubenswrapper[4980]: E1206 03:34:23.236170 4980 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 03:34:23 crc kubenswrapper[4980]: I1206 03:34:23.239235 4980 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="01ab3dd5-8196-46d0-ad33-122e2ca51def" path="/var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes" Dec 06 03:34:23 crc kubenswrapper[4980]: I1206 03:34:23.240019 4980 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" path="/var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes" Dec 06 03:34:23 crc kubenswrapper[4980]: I1206 03:34:23.241661 4980 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09efc573-dbb6-4249-bd59-9b87aba8dd28" path="/var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes" Dec 06 03:34:23 crc kubenswrapper[4980]: I1206 03:34:23.242292 4980 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b574797-001e-440a-8f4e-c0be86edad0f" path="/var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes" Dec 06 03:34:23 crc kubenswrapper[4980]: I1206 03:34:23.244135 4980 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b78653f-4ff9-4508-8672-245ed9b561e3" path="/var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes" Dec 06 03:34:23 crc kubenswrapper[4980]: I1206 03:34:23.244752 4980 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1386a44e-36a2-460c-96d0-0359d2b6f0f5" path="/var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes" Dec 06 03:34:23 crc kubenswrapper[4980]: I1206 03:34:23.245564 4980 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1bf7eb37-55a3-4c65-b768-a94c82151e69" path="/var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes" Dec 06 03:34:23 crc kubenswrapper[4980]: I1206 03:34:23.247761 4980 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1d611f23-29be-4491-8495-bee1670e935f" path="/var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes" Dec 06 03:34:23 crc kubenswrapper[4980]: I1206 03:34:23.248399 4980 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="20b0d48f-5fd6-431c-a545-e3c800c7b866" path="/var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/volumes" Dec 06 03:34:23 crc kubenswrapper[4980]: I1206 03:34:23.249834 4980 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" path="/var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes" Dec 06 03:34:23 crc kubenswrapper[4980]: I1206 03:34:23.250546 4980 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="22c825df-677d-4ca6-82db-3454ed06e783" path="/var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes" Dec 06 03:34:23 crc kubenswrapper[4980]: I1206 03:34:23.251876 4980 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="25e176fe-21b4-4974-b1ed-c8b94f112a7f" path="/var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes" Dec 06 03:34:23 crc kubenswrapper[4980]: I1206 03:34:23.252448 4980 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" path="/var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes" Dec 06 03:34:23 crc kubenswrapper[4980]: I1206 03:34:23.254447 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:22Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 06 03:34:23 crc kubenswrapper[4980]: I1206 03:34:23.256113 4980 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="31d8b7a1-420e-4252-a5b7-eebe8a111292" path="/var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes" Dec 06 03:34:23 crc kubenswrapper[4980]: I1206 03:34:23.256945 4980 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3ab1a177-2de0-46d9-b765-d0d0649bb42e" path="/var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/volumes" Dec 06 03:34:23 crc kubenswrapper[4980]: I1206 03:34:23.258219 4980 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" path="/var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes" Dec 06 03:34:23 crc kubenswrapper[4980]: I1206 03:34:23.258947 4980 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="43509403-f426-496e-be36-56cef71462f5" path="/var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes" Dec 06 03:34:23 crc kubenswrapper[4980]: I1206 03:34:23.259403 4980 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="44663579-783b-4372-86d6-acf235a62d72" path="/var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/volumes" Dec 06 03:34:23 crc kubenswrapper[4980]: I1206 03:34:23.260430 4980 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="496e6271-fb68-4057-954e-a0d97a4afa3f" path="/var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes" Dec 06 03:34:23 crc kubenswrapper[4980]: I1206 03:34:23.261076 4980 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" path="/var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes" Dec 06 03:34:23 crc kubenswrapper[4980]: I1206 03:34:23.261694 4980 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49ef4625-1d3a-4a9f-b595-c2433d32326d" path="/var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/volumes" Dec 06 03:34:23 crc kubenswrapper[4980]: I1206 03:34:23.262861 4980 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4bb40260-dbaa-4fb0-84df-5e680505d512" path="/var/lib/kubelet/pods/4bb40260-dbaa-4fb0-84df-5e680505d512/volumes" Dec 06 03:34:23 crc kubenswrapper[4980]: I1206 03:34:23.263337 4980 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5225d0e4-402f-4861-b410-819f433b1803" path="/var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes" Dec 06 03:34:23 crc kubenswrapper[4980]: I1206 03:34:23.300746 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:22Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 06 03:34:23 crc kubenswrapper[4980]: I1206 03:34:23.313850 4980 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5441d097-087c-4d9a-baa8-b210afa90fc9" path="/var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes" Dec 06 03:34:23 crc kubenswrapper[4980]: I1206 03:34:23.314432 4980 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="57a731c4-ef35-47a8-b875-bfb08a7f8011" path="/var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes" Dec 06 03:34:23 crc kubenswrapper[4980]: I1206 03:34:23.316448 4980 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5b88f790-22fa-440e-b583-365168c0b23d" path="/var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/volumes" Dec 06 03:34:23 crc kubenswrapper[4980]: I1206 03:34:23.317426 4980 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5fe579f8-e8a6-4643-bce5-a661393c4dde" path="/var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/volumes" Dec 06 03:34:23 crc kubenswrapper[4980]: I1206 03:34:23.351990 4980 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6402fda4-df10-493c-b4e5-d0569419652d" path="/var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes" Dec 06 03:34:23 crc kubenswrapper[4980]: I1206 03:34:23.360148 4980 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6509e943-70c6-444c-bc41-48a544e36fbd" path="/var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes" Dec 06 03:34:23 crc kubenswrapper[4980]: I1206 03:34:23.366350 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 06 03:34:23 crc kubenswrapper[4980]: I1206 03:34:23.388466 4980 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6731426b-95fe-49ff-bb5f-40441049fde2" path="/var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/volumes" Dec 06 03:34:23 crc kubenswrapper[4980]: I1206 03:34:23.388994 4980 kubelet_volumes.go:152] "Cleaned up orphaned volume subpath from pod" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volume-subpaths/run-systemd/ovnkube-controller/6" Dec 06 03:34:23 crc kubenswrapper[4980]: I1206 03:34:23.389099 4980 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volumes" Dec 06 03:34:23 crc kubenswrapper[4980]: I1206 03:34:23.407637 4980 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Dec 06 03:34:23 crc kubenswrapper[4980]: I1206 03:34:23.417523 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 06 03:34:23 crc kubenswrapper[4980]: I1206 03:34:23.462865 4980 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7539238d-5fe0-46ed-884e-1c3b566537ec" path="/var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes" Dec 06 03:34:23 crc kubenswrapper[4980]: I1206 03:34:23.463667 4980 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7583ce53-e0fe-4a16-9e4d-50516596a136" path="/var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes" Dec 06 03:34:23 crc kubenswrapper[4980]: I1206 03:34:23.464228 4980 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7bb08738-c794-4ee8-9972-3a62ca171029" path="/var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes" Dec 06 03:34:23 crc kubenswrapper[4980]: I1206 03:34:23.467568 4980 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="87cf06ed-a83f-41a7-828d-70653580a8cb" path="/var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes" Dec 06 03:34:23 crc kubenswrapper[4980]: I1206 03:34:23.496911 4980 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" path="/var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes" Dec 06 03:34:23 crc kubenswrapper[4980]: I1206 03:34:23.497619 4980 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="925f1c65-6136-48ba-85aa-3a3b50560753" path="/var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes" Dec 06 03:34:23 crc kubenswrapper[4980]: I1206 03:34:23.498331 4980 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" path="/var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/volumes" Dec 06 03:34:23 crc kubenswrapper[4980]: I1206 03:34:23.499087 4980 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9d4552c7-cd75-42dd-8880-30dd377c49a4" path="/var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes" Dec 06 03:34:23 crc kubenswrapper[4980]: I1206 03:34:23.499648 4980 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" path="/var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/volumes" Dec 06 03:34:23 crc kubenswrapper[4980]: I1206 03:34:23.500289 4980 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a31745f5-9847-4afe-82a5-3161cc66ca93" path="/var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes" Dec 06 03:34:23 crc kubenswrapper[4980]: I1206 03:34:23.513161 4980 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" path="/var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes" Dec 06 03:34:23 crc kubenswrapper[4980]: I1206 03:34:23.514389 4980 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6312bbd-5731-4ea0-a20f-81d5a57df44a" path="/var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/volumes" Dec 06 03:34:23 crc kubenswrapper[4980]: I1206 03:34:23.515005 4980 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" path="/var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes" Dec 06 03:34:23 crc kubenswrapper[4980]: I1206 03:34:23.515670 4980 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" path="/var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes" Dec 06 03:34:23 crc kubenswrapper[4980]: I1206 03:34:23.516755 4980 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" path="/var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/volumes" Dec 06 03:34:23 crc kubenswrapper[4980]: I1206 03:34:23.517664 4980 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bf126b07-da06-4140-9a57-dfd54fc6b486" path="/var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes" Dec 06 03:34:23 crc kubenswrapper[4980]: I1206 03:34:23.518536 4980 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c03ee662-fb2f-4fc4-a2c1-af487c19d254" path="/var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes" Dec 06 03:34:23 crc kubenswrapper[4980]: I1206 03:34:23.519033 4980 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" path="/var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/volumes" Dec 06 03:34:23 crc kubenswrapper[4980]: I1206 03:34:23.519920 4980 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e7e6199b-1264-4501-8953-767f51328d08" path="/var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes" Dec 06 03:34:23 crc kubenswrapper[4980]: I1206 03:34:23.520432 4980 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="efdd0498-1daa-4136-9a4a-3b948c2293fc" path="/var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/volumes" Dec 06 03:34:23 crc kubenswrapper[4980]: I1206 03:34:23.521144 4980 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" path="/var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/volumes" Dec 06 03:34:23 crc kubenswrapper[4980]: I1206 03:34:23.522047 4980 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fda69060-fa79-4696-b1a6-7980f124bf7c" path="/var/lib/kubelet/pods/fda69060-fa79-4696-b1a6-7980f124bf7c/volumes" Dec 06 03:34:23 crc kubenswrapper[4980]: I1206 03:34:23.522639 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"93a1a57b5a61b16f2bfc47b2e745874afa0848bb8e74ef6832c3c8b6ed85bf3a"} Dec 06 03:34:23 crc kubenswrapper[4980]: I1206 03:34:23.522685 4980 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 06 03:34:23 crc kubenswrapper[4980]: I1206 03:34:23.522703 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"334ea6a273fd7d268906c9f4a6dbf6f4df4fcdbf38f3d7d248cb7d6cf8537794"} Dec 06 03:34:23 crc kubenswrapper[4980]: I1206 03:34:23.522723 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"4ec44dcd5e4bf0a0466bb586ebd61259364dbbfe22499f4cbff854e3bd153082"} Dec 06 03:34:23 crc kubenswrapper[4980]: I1206 03:34:23.522777 4980 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 06 03:34:23 crc kubenswrapper[4980]: I1206 03:34:23.522790 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"cb8e711cd9f04199553918c742ba12a89d24f07c2d55bbc2e60e379ef2f83428"} Dec 06 03:34:23 crc kubenswrapper[4980]: I1206 03:34:23.522800 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"bafaa9670b05671fb57b658c2960bf3ee32c84e52f86a3a3bbab01d5a38ff0c4"} Dec 06 03:34:23 crc kubenswrapper[4980]: I1206 03:34:23.522811 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"8b794c12da7f5589be20c736f2c23af29c1a94ade66fbd510b09b9bc19b2af45"} Dec 06 03:34:23 crc kubenswrapper[4980]: I1206 03:34:23.523580 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:22Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 06 03:34:23 crc kubenswrapper[4980]: I1206 03:34:23.532680 4980 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/node-resolver-sv5sz"] Dec 06 03:34:23 crc kubenswrapper[4980]: I1206 03:34:23.533026 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-sv5sz" Dec 06 03:34:23 crc kubenswrapper[4980]: I1206 03:34:23.533943 4980 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-additional-cni-plugins-dx44q"] Dec 06 03:34:23 crc kubenswrapper[4980]: I1206 03:34:23.534548 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-dx44q" Dec 06 03:34:23 crc kubenswrapper[4980]: I1206 03:34:23.546787 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/e5bac22a-512c-49d0-90db-47de6aaaefd0-cni-binary-copy\") pod \"multus-additional-cni-plugins-dx44q\" (UID: \"e5bac22a-512c-49d0-90db-47de6aaaefd0\") " pod="openshift-multus/multus-additional-cni-plugins-dx44q" Dec 06 03:34:23 crc kubenswrapper[4980]: I1206 03:34:23.546851 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/e5bac22a-512c-49d0-90db-47de6aaaefd0-os-release\") pod \"multus-additional-cni-plugins-dx44q\" (UID: \"e5bac22a-512c-49d0-90db-47de6aaaefd0\") " pod="openshift-multus/multus-additional-cni-plugins-dx44q" Dec 06 03:34:23 crc kubenswrapper[4980]: I1206 03:34:23.546874 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/873977aa-4eb0-4526-bc0b-757ad2dea2a5-hosts-file\") pod \"node-resolver-sv5sz\" (UID: \"873977aa-4eb0-4526-bc0b-757ad2dea2a5\") " pod="openshift-dns/node-resolver-sv5sz" Dec 06 03:34:23 crc kubenswrapper[4980]: I1206 03:34:23.546894 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lg7tx\" (UniqueName: \"kubernetes.io/projected/873977aa-4eb0-4526-bc0b-757ad2dea2a5-kube-api-access-lg7tx\") pod \"node-resolver-sv5sz\" (UID: \"873977aa-4eb0-4526-bc0b-757ad2dea2a5\") " pod="openshift-dns/node-resolver-sv5sz" Dec 06 03:34:23 crc kubenswrapper[4980]: I1206 03:34:23.546924 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/e5bac22a-512c-49d0-90db-47de6aaaefd0-tuning-conf-dir\") pod \"multus-additional-cni-plugins-dx44q\" (UID: \"e5bac22a-512c-49d0-90db-47de6aaaefd0\") " pod="openshift-multus/multus-additional-cni-plugins-dx44q" Dec 06 03:34:23 crc kubenswrapper[4980]: I1206 03:34:23.546944 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/e5bac22a-512c-49d0-90db-47de6aaaefd0-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-dx44q\" (UID: \"e5bac22a-512c-49d0-90db-47de6aaaefd0\") " pod="openshift-multus/multus-additional-cni-plugins-dx44q" Dec 06 03:34:23 crc kubenswrapper[4980]: I1206 03:34:23.546976 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f7mdm\" (UniqueName: \"kubernetes.io/projected/e5bac22a-512c-49d0-90db-47de6aaaefd0-kube-api-access-f7mdm\") pod \"multus-additional-cni-plugins-dx44q\" (UID: \"e5bac22a-512c-49d0-90db-47de6aaaefd0\") " pod="openshift-multus/multus-additional-cni-plugins-dx44q" Dec 06 03:34:23 crc kubenswrapper[4980]: I1206 03:34:23.546997 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/e5bac22a-512c-49d0-90db-47de6aaaefd0-system-cni-dir\") pod \"multus-additional-cni-plugins-dx44q\" (UID: \"e5bac22a-512c-49d0-90db-47de6aaaefd0\") " pod="openshift-multus/multus-additional-cni-plugins-dx44q" Dec 06 03:34:23 crc kubenswrapper[4980]: I1206 03:34:23.547016 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/e5bac22a-512c-49d0-90db-47de6aaaefd0-cnibin\") pod \"multus-additional-cni-plugins-dx44q\" (UID: \"e5bac22a-512c-49d0-90db-47de6aaaefd0\") " pod="openshift-multus/multus-additional-cni-plugins-dx44q" Dec 06 03:34:23 crc kubenswrapper[4980]: I1206 03:34:23.577664 4980 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-kq4z8"] Dec 06 03:34:23 crc kubenswrapper[4980]: I1206 03:34:23.577886 4980 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Dec 06 03:34:23 crc kubenswrapper[4980]: I1206 03:34:23.578126 4980 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Dec 06 03:34:23 crc kubenswrapper[4980]: I1206 03:34:23.578330 4980 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Dec 06 03:34:23 crc kubenswrapper[4980]: I1206 03:34:23.578470 4980 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Dec 06 03:34:23 crc kubenswrapper[4980]: I1206 03:34:23.578618 4980 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Dec 06 03:34:23 crc kubenswrapper[4980]: I1206 03:34:23.578635 4980 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Dec 06 03:34:23 crc kubenswrapper[4980]: I1206 03:34:23.578851 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-kq4z8" Dec 06 03:34:23 crc kubenswrapper[4980]: I1206 03:34:23.578912 4980 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Dec 06 03:34:23 crc kubenswrapper[4980]: I1206 03:34:23.578969 4980 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Dec 06 03:34:23 crc kubenswrapper[4980]: I1206 03:34:23.579012 4980 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-5w4fr"] Dec 06 03:34:23 crc kubenswrapper[4980]: I1206 03:34:23.582369 4980 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Dec 06 03:34:23 crc kubenswrapper[4980]: I1206 03:34:23.582684 4980 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Dec 06 03:34:23 crc kubenswrapper[4980]: I1206 03:34:23.584311 4980 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-daemon-r5zfb"] Dec 06 03:34:23 crc kubenswrapper[4980]: I1206 03:34:23.584696 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-5w4fr" Dec 06 03:34:23 crc kubenswrapper[4980]: I1206 03:34:23.602925 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:22Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 06 03:34:23 crc kubenswrapper[4980]: I1206 03:34:23.603194 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-r5zfb" Dec 06 03:34:23 crc kubenswrapper[4980]: I1206 03:34:23.603491 4980 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Dec 06 03:34:23 crc kubenswrapper[4980]: I1206 03:34:23.603533 4980 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Dec 06 03:34:23 crc kubenswrapper[4980]: I1206 03:34:23.604043 4980 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Dec 06 03:34:23 crc kubenswrapper[4980]: I1206 03:34:23.604075 4980 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Dec 06 03:34:23 crc kubenswrapper[4980]: I1206 03:34:23.605016 4980 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Dec 06 03:34:23 crc kubenswrapper[4980]: I1206 03:34:23.605335 4980 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Dec 06 03:34:23 crc kubenswrapper[4980]: I1206 03:34:23.610877 4980 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Dec 06 03:34:23 crc kubenswrapper[4980]: I1206 03:34:23.616456 4980 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Dec 06 03:34:23 crc kubenswrapper[4980]: I1206 03:34:23.617208 4980 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Dec 06 03:34:23 crc kubenswrapper[4980]: I1206 03:34:23.617356 4980 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Dec 06 03:34:23 crc kubenswrapper[4980]: I1206 03:34:23.617469 4980 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Dec 06 03:34:23 crc kubenswrapper[4980]: I1206 03:34:23.619316 4980 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Dec 06 03:34:23 crc kubenswrapper[4980]: I1206 03:34:23.645665 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cab608b3-a02f-475b-9aae-9b8f5d878de9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f7fcc44b42bcbb55509048e3a566bb557333c4a83e33d7b0fba8eb514317909\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c22e7dd8886c2dfbb742900ed86ec7b0c2edf9a1363e74f6666961d928f06c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://473e792b2d0a9644854ee0581c6120fc9c989bc342899ef72b1d056080fd4694\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ee8e936dfc39361d6f99460910591ca8c901a20f0d274bbb6aaebfa8a00ee06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://77bc12768b018625ba11a6bfd66688f926c4b7f3bc1d0e41c6551a18b485ecf3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://27cf07cd96893900da4afecd0d56fd9269bc60c1d23246b4685fcd8756357c19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://27cf07cd96893900da4afecd0d56fd9269bc60c1d23246b4685fcd8756357c19\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c49feb24566a8284378670fa0390710d38c237acc94ffb654f90f3fa4626e4c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c49feb24566a8284378670fa0390710d38c237acc94ffb654f90f3fa4626e4c8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:03Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://bb74d1c3d752a379fa394c450217badb45511fbcafad62f7b618131cab94b29e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bb74d1c3d752a379fa394c450217badb45511fbcafad62f7b618131cab94b29e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:01Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 06 03:34:23 crc kubenswrapper[4980]: I1206 03:34:23.647448 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/5af0e768-3c9c-4401-ab56-57516cd3170e-node-log\") pod \"ovnkube-node-5w4fr\" (UID: \"5af0e768-3c9c-4401-ab56-57516cd3170e\") " pod="openshift-ovn-kubernetes/ovnkube-node-5w4fr" Dec 06 03:34:23 crc kubenswrapper[4980]: I1206 03:34:23.647481 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/5af0e768-3c9c-4401-ab56-57516cd3170e-run-openvswitch\") pod \"ovnkube-node-5w4fr\" (UID: \"5af0e768-3c9c-4401-ab56-57516cd3170e\") " pod="openshift-ovn-kubernetes/ovnkube-node-5w4fr" Dec 06 03:34:23 crc kubenswrapper[4980]: I1206 03:34:23.647497 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/5af0e768-3c9c-4401-ab56-57516cd3170e-host-run-ovn-kubernetes\") pod \"ovnkube-node-5w4fr\" (UID: \"5af0e768-3c9c-4401-ab56-57516cd3170e\") " pod="openshift-ovn-kubernetes/ovnkube-node-5w4fr" Dec 06 03:34:23 crc kubenswrapper[4980]: I1206 03:34:23.647529 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/f96c1f6c-55e8-4d86-bff3-5381581b7022-cni-binary-copy\") pod \"multus-kq4z8\" (UID: \"f96c1f6c-55e8-4d86-bff3-5381581b7022\") " pod="openshift-multus/multus-kq4z8" Dec 06 03:34:23 crc kubenswrapper[4980]: I1206 03:34:23.647543 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/5af0e768-3c9c-4401-ab56-57516cd3170e-etc-openvswitch\") pod \"ovnkube-node-5w4fr\" (UID: \"5af0e768-3c9c-4401-ab56-57516cd3170e\") " pod="openshift-ovn-kubernetes/ovnkube-node-5w4fr" Dec 06 03:34:23 crc kubenswrapper[4980]: I1206 03:34:23.647563 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/e5bac22a-512c-49d0-90db-47de6aaaefd0-os-release\") pod \"multus-additional-cni-plugins-dx44q\" (UID: \"e5bac22a-512c-49d0-90db-47de6aaaefd0\") " pod="openshift-multus/multus-additional-cni-plugins-dx44q" Dec 06 03:34:23 crc kubenswrapper[4980]: I1206 03:34:23.647578 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/873977aa-4eb0-4526-bc0b-757ad2dea2a5-hosts-file\") pod \"node-resolver-sv5sz\" (UID: \"873977aa-4eb0-4526-bc0b-757ad2dea2a5\") " pod="openshift-dns/node-resolver-sv5sz" Dec 06 03:34:23 crc kubenswrapper[4980]: I1206 03:34:23.647593 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lg7tx\" (UniqueName: \"kubernetes.io/projected/873977aa-4eb0-4526-bc0b-757ad2dea2a5-kube-api-access-lg7tx\") pod \"node-resolver-sv5sz\" (UID: \"873977aa-4eb0-4526-bc0b-757ad2dea2a5\") " pod="openshift-dns/node-resolver-sv5sz" Dec 06 03:34:23 crc kubenswrapper[4980]: I1206 03:34:23.647610 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zhm5t\" (UniqueName: \"kubernetes.io/projected/320f44d1-a671-4a91-b328-a8b0fdd8f23a-kube-api-access-zhm5t\") pod \"machine-config-daemon-r5zfb\" (UID: \"320f44d1-a671-4a91-b328-a8b0fdd8f23a\") " pod="openshift-machine-config-operator/machine-config-daemon-r5zfb" Dec 06 03:34:23 crc kubenswrapper[4980]: I1206 03:34:23.647624 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/5af0e768-3c9c-4401-ab56-57516cd3170e-run-systemd\") pod \"ovnkube-node-5w4fr\" (UID: \"5af0e768-3c9c-4401-ab56-57516cd3170e\") " pod="openshift-ovn-kubernetes/ovnkube-node-5w4fr" Dec 06 03:34:23 crc kubenswrapper[4980]: I1206 03:34:23.647638 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/f96c1f6c-55e8-4d86-bff3-5381581b7022-cnibin\") pod \"multus-kq4z8\" (UID: \"f96c1f6c-55e8-4d86-bff3-5381581b7022\") " pod="openshift-multus/multus-kq4z8" Dec 06 03:34:23 crc kubenswrapper[4980]: I1206 03:34:23.647655 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/f96c1f6c-55e8-4d86-bff3-5381581b7022-host-var-lib-cni-bin\") pod \"multus-kq4z8\" (UID: \"f96c1f6c-55e8-4d86-bff3-5381581b7022\") " pod="openshift-multus/multus-kq4z8" Dec 06 03:34:23 crc kubenswrapper[4980]: I1206 03:34:23.647668 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/320f44d1-a671-4a91-b328-a8b0fdd8f23a-rootfs\") pod \"machine-config-daemon-r5zfb\" (UID: \"320f44d1-a671-4a91-b328-a8b0fdd8f23a\") " pod="openshift-machine-config-operator/machine-config-daemon-r5zfb" Dec 06 03:34:23 crc kubenswrapper[4980]: I1206 03:34:23.647690 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/f96c1f6c-55e8-4d86-bff3-5381581b7022-etc-kubernetes\") pod \"multus-kq4z8\" (UID: \"f96c1f6c-55e8-4d86-bff3-5381581b7022\") " pod="openshift-multus/multus-kq4z8" Dec 06 03:34:23 crc kubenswrapper[4980]: I1206 03:34:23.647706 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/5af0e768-3c9c-4401-ab56-57516cd3170e-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-5w4fr\" (UID: \"5af0e768-3c9c-4401-ab56-57516cd3170e\") " pod="openshift-ovn-kubernetes/ovnkube-node-5w4fr" Dec 06 03:34:23 crc kubenswrapper[4980]: I1206 03:34:23.647722 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/e5bac22a-512c-49d0-90db-47de6aaaefd0-system-cni-dir\") pod \"multus-additional-cni-plugins-dx44q\" (UID: \"e5bac22a-512c-49d0-90db-47de6aaaefd0\") " pod="openshift-multus/multus-additional-cni-plugins-dx44q" Dec 06 03:34:23 crc kubenswrapper[4980]: I1206 03:34:23.647737 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/e5bac22a-512c-49d0-90db-47de6aaaefd0-cnibin\") pod \"multus-additional-cni-plugins-dx44q\" (UID: \"e5bac22a-512c-49d0-90db-47de6aaaefd0\") " pod="openshift-multus/multus-additional-cni-plugins-dx44q" Dec 06 03:34:23 crc kubenswrapper[4980]: I1206 03:34:23.647752 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/f96c1f6c-55e8-4d86-bff3-5381581b7022-hostroot\") pod \"multus-kq4z8\" (UID: \"f96c1f6c-55e8-4d86-bff3-5381581b7022\") " pod="openshift-multus/multus-kq4z8" Dec 06 03:34:23 crc kubenswrapper[4980]: I1206 03:34:23.647770 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/320f44d1-a671-4a91-b328-a8b0fdd8f23a-mcd-auth-proxy-config\") pod \"machine-config-daemon-r5zfb\" (UID: \"320f44d1-a671-4a91-b328-a8b0fdd8f23a\") " pod="openshift-machine-config-operator/machine-config-daemon-r5zfb" Dec 06 03:34:23 crc kubenswrapper[4980]: I1206 03:34:23.647785 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/5af0e768-3c9c-4401-ab56-57516cd3170e-var-lib-openvswitch\") pod \"ovnkube-node-5w4fr\" (UID: \"5af0e768-3c9c-4401-ab56-57516cd3170e\") " pod="openshift-ovn-kubernetes/ovnkube-node-5w4fr" Dec 06 03:34:23 crc kubenswrapper[4980]: I1206 03:34:23.647798 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/5af0e768-3c9c-4401-ab56-57516cd3170e-ovnkube-config\") pod \"ovnkube-node-5w4fr\" (UID: \"5af0e768-3c9c-4401-ab56-57516cd3170e\") " pod="openshift-ovn-kubernetes/ovnkube-node-5w4fr" Dec 06 03:34:23 crc kubenswrapper[4980]: I1206 03:34:23.647812 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bdx7g\" (UniqueName: \"kubernetes.io/projected/f96c1f6c-55e8-4d86-bff3-5381581b7022-kube-api-access-bdx7g\") pod \"multus-kq4z8\" (UID: \"f96c1f6c-55e8-4d86-bff3-5381581b7022\") " pod="openshift-multus/multus-kq4z8" Dec 06 03:34:23 crc kubenswrapper[4980]: I1206 03:34:23.647951 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/5af0e768-3c9c-4401-ab56-57516cd3170e-run-ovn\") pod \"ovnkube-node-5w4fr\" (UID: \"5af0e768-3c9c-4401-ab56-57516cd3170e\") " pod="openshift-ovn-kubernetes/ovnkube-node-5w4fr" Dec 06 03:34:23 crc kubenswrapper[4980]: I1206 03:34:23.647987 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/e5bac22a-512c-49d0-90db-47de6aaaefd0-cni-binary-copy\") pod \"multus-additional-cni-plugins-dx44q\" (UID: \"e5bac22a-512c-49d0-90db-47de6aaaefd0\") " pod="openshift-multus/multus-additional-cni-plugins-dx44q" Dec 06 03:34:23 crc kubenswrapper[4980]: I1206 03:34:23.648001 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/f96c1f6c-55e8-4d86-bff3-5381581b7022-os-release\") pod \"multus-kq4z8\" (UID: \"f96c1f6c-55e8-4d86-bff3-5381581b7022\") " pod="openshift-multus/multus-kq4z8" Dec 06 03:34:23 crc kubenswrapper[4980]: I1206 03:34:23.648031 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/f96c1f6c-55e8-4d86-bff3-5381581b7022-multus-socket-dir-parent\") pod \"multus-kq4z8\" (UID: \"f96c1f6c-55e8-4d86-bff3-5381581b7022\") " pod="openshift-multus/multus-kq4z8" Dec 06 03:34:23 crc kubenswrapper[4980]: I1206 03:34:23.648060 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/f96c1f6c-55e8-4d86-bff3-5381581b7022-host-run-multus-certs\") pod \"multus-kq4z8\" (UID: \"f96c1f6c-55e8-4d86-bff3-5381581b7022\") " pod="openshift-multus/multus-kq4z8" Dec 06 03:34:23 crc kubenswrapper[4980]: I1206 03:34:23.648075 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/f96c1f6c-55e8-4d86-bff3-5381581b7022-host-run-k8s-cni-cncf-io\") pod \"multus-kq4z8\" (UID: \"f96c1f6c-55e8-4d86-bff3-5381581b7022\") " pod="openshift-multus/multus-kq4z8" Dec 06 03:34:23 crc kubenswrapper[4980]: I1206 03:34:23.648099 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/5af0e768-3c9c-4401-ab56-57516cd3170e-systemd-units\") pod \"ovnkube-node-5w4fr\" (UID: \"5af0e768-3c9c-4401-ab56-57516cd3170e\") " pod="openshift-ovn-kubernetes/ovnkube-node-5w4fr" Dec 06 03:34:23 crc kubenswrapper[4980]: I1206 03:34:23.648116 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/5af0e768-3c9c-4401-ab56-57516cd3170e-host-run-netns\") pod \"ovnkube-node-5w4fr\" (UID: \"5af0e768-3c9c-4401-ab56-57516cd3170e\") " pod="openshift-ovn-kubernetes/ovnkube-node-5w4fr" Dec 06 03:34:23 crc kubenswrapper[4980]: I1206 03:34:23.648161 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zfzrb\" (UniqueName: \"kubernetes.io/projected/5af0e768-3c9c-4401-ab56-57516cd3170e-kube-api-access-zfzrb\") pod \"ovnkube-node-5w4fr\" (UID: \"5af0e768-3c9c-4401-ab56-57516cd3170e\") " pod="openshift-ovn-kubernetes/ovnkube-node-5w4fr" Dec 06 03:34:23 crc kubenswrapper[4980]: I1206 03:34:23.648177 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/5af0e768-3c9c-4401-ab56-57516cd3170e-env-overrides\") pod \"ovnkube-node-5w4fr\" (UID: \"5af0e768-3c9c-4401-ab56-57516cd3170e\") " pod="openshift-ovn-kubernetes/ovnkube-node-5w4fr" Dec 06 03:34:23 crc kubenswrapper[4980]: I1206 03:34:23.648192 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/5af0e768-3c9c-4401-ab56-57516cd3170e-ovn-node-metrics-cert\") pod \"ovnkube-node-5w4fr\" (UID: \"5af0e768-3c9c-4401-ab56-57516cd3170e\") " pod="openshift-ovn-kubernetes/ovnkube-node-5w4fr" Dec 06 03:34:23 crc kubenswrapper[4980]: I1206 03:34:23.648228 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/e5bac22a-512c-49d0-90db-47de6aaaefd0-tuning-conf-dir\") pod \"multus-additional-cni-plugins-dx44q\" (UID: \"e5bac22a-512c-49d0-90db-47de6aaaefd0\") " pod="openshift-multus/multus-additional-cni-plugins-dx44q" Dec 06 03:34:23 crc kubenswrapper[4980]: I1206 03:34:23.648248 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/e5bac22a-512c-49d0-90db-47de6aaaefd0-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-dx44q\" (UID: \"e5bac22a-512c-49d0-90db-47de6aaaefd0\") " pod="openshift-multus/multus-additional-cni-plugins-dx44q" Dec 06 03:34:23 crc kubenswrapper[4980]: I1206 03:34:23.648263 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/f96c1f6c-55e8-4d86-bff3-5381581b7022-multus-cni-dir\") pod \"multus-kq4z8\" (UID: \"f96c1f6c-55e8-4d86-bff3-5381581b7022\") " pod="openshift-multus/multus-kq4z8" Dec 06 03:34:23 crc kubenswrapper[4980]: I1206 03:34:23.648280 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/f96c1f6c-55e8-4d86-bff3-5381581b7022-multus-daemon-config\") pod \"multus-kq4z8\" (UID: \"f96c1f6c-55e8-4d86-bff3-5381581b7022\") " pod="openshift-multus/multus-kq4z8" Dec 06 03:34:23 crc kubenswrapper[4980]: I1206 03:34:23.648297 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/5af0e768-3c9c-4401-ab56-57516cd3170e-host-cni-netd\") pod \"ovnkube-node-5w4fr\" (UID: \"5af0e768-3c9c-4401-ab56-57516cd3170e\") " pod="openshift-ovn-kubernetes/ovnkube-node-5w4fr" Dec 06 03:34:23 crc kubenswrapper[4980]: I1206 03:34:23.648313 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f7mdm\" (UniqueName: \"kubernetes.io/projected/e5bac22a-512c-49d0-90db-47de6aaaefd0-kube-api-access-f7mdm\") pod \"multus-additional-cni-plugins-dx44q\" (UID: \"e5bac22a-512c-49d0-90db-47de6aaaefd0\") " pod="openshift-multus/multus-additional-cni-plugins-dx44q" Dec 06 03:34:23 crc kubenswrapper[4980]: I1206 03:34:23.648330 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/f96c1f6c-55e8-4d86-bff3-5381581b7022-system-cni-dir\") pod \"multus-kq4z8\" (UID: \"f96c1f6c-55e8-4d86-bff3-5381581b7022\") " pod="openshift-multus/multus-kq4z8" Dec 06 03:34:23 crc kubenswrapper[4980]: I1206 03:34:23.648345 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/f96c1f6c-55e8-4d86-bff3-5381581b7022-host-run-netns\") pod \"multus-kq4z8\" (UID: \"f96c1f6c-55e8-4d86-bff3-5381581b7022\") " pod="openshift-multus/multus-kq4z8" Dec 06 03:34:23 crc kubenswrapper[4980]: I1206 03:34:23.648360 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/320f44d1-a671-4a91-b328-a8b0fdd8f23a-proxy-tls\") pod \"machine-config-daemon-r5zfb\" (UID: \"320f44d1-a671-4a91-b328-a8b0fdd8f23a\") " pod="openshift-machine-config-operator/machine-config-daemon-r5zfb" Dec 06 03:34:23 crc kubenswrapper[4980]: I1206 03:34:23.648378 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/5af0e768-3c9c-4401-ab56-57516cd3170e-log-socket\") pod \"ovnkube-node-5w4fr\" (UID: \"5af0e768-3c9c-4401-ab56-57516cd3170e\") " pod="openshift-ovn-kubernetes/ovnkube-node-5w4fr" Dec 06 03:34:23 crc kubenswrapper[4980]: I1206 03:34:23.648394 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/f96c1f6c-55e8-4d86-bff3-5381581b7022-host-var-lib-cni-multus\") pod \"multus-kq4z8\" (UID: \"f96c1f6c-55e8-4d86-bff3-5381581b7022\") " pod="openshift-multus/multus-kq4z8" Dec 06 03:34:23 crc kubenswrapper[4980]: I1206 03:34:23.648422 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/5af0e768-3c9c-4401-ab56-57516cd3170e-host-kubelet\") pod \"ovnkube-node-5w4fr\" (UID: \"5af0e768-3c9c-4401-ab56-57516cd3170e\") " pod="openshift-ovn-kubernetes/ovnkube-node-5w4fr" Dec 06 03:34:23 crc kubenswrapper[4980]: I1206 03:34:23.648436 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/5af0e768-3c9c-4401-ab56-57516cd3170e-ovnkube-script-lib\") pod \"ovnkube-node-5w4fr\" (UID: \"5af0e768-3c9c-4401-ab56-57516cd3170e\") " pod="openshift-ovn-kubernetes/ovnkube-node-5w4fr" Dec 06 03:34:23 crc kubenswrapper[4980]: I1206 03:34:23.648451 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/f96c1f6c-55e8-4d86-bff3-5381581b7022-host-var-lib-kubelet\") pod \"multus-kq4z8\" (UID: \"f96c1f6c-55e8-4d86-bff3-5381581b7022\") " pod="openshift-multus/multus-kq4z8" Dec 06 03:34:23 crc kubenswrapper[4980]: I1206 03:34:23.648466 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/5af0e768-3c9c-4401-ab56-57516cd3170e-host-slash\") pod \"ovnkube-node-5w4fr\" (UID: \"5af0e768-3c9c-4401-ab56-57516cd3170e\") " pod="openshift-ovn-kubernetes/ovnkube-node-5w4fr" Dec 06 03:34:23 crc kubenswrapper[4980]: I1206 03:34:23.648488 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/f96c1f6c-55e8-4d86-bff3-5381581b7022-multus-conf-dir\") pod \"multus-kq4z8\" (UID: \"f96c1f6c-55e8-4d86-bff3-5381581b7022\") " pod="openshift-multus/multus-kq4z8" Dec 06 03:34:23 crc kubenswrapper[4980]: I1206 03:34:23.648503 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/5af0e768-3c9c-4401-ab56-57516cd3170e-host-cni-bin\") pod \"ovnkube-node-5w4fr\" (UID: \"5af0e768-3c9c-4401-ab56-57516cd3170e\") " pod="openshift-ovn-kubernetes/ovnkube-node-5w4fr" Dec 06 03:34:23 crc kubenswrapper[4980]: I1206 03:34:23.648735 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/e5bac22a-512c-49d0-90db-47de6aaaefd0-os-release\") pod \"multus-additional-cni-plugins-dx44q\" (UID: \"e5bac22a-512c-49d0-90db-47de6aaaefd0\") " pod="openshift-multus/multus-additional-cni-plugins-dx44q" Dec 06 03:34:23 crc kubenswrapper[4980]: I1206 03:34:23.648780 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/873977aa-4eb0-4526-bc0b-757ad2dea2a5-hosts-file\") pod \"node-resolver-sv5sz\" (UID: \"873977aa-4eb0-4526-bc0b-757ad2dea2a5\") " pod="openshift-dns/node-resolver-sv5sz" Dec 06 03:34:23 crc kubenswrapper[4980]: I1206 03:34:23.649008 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/e5bac22a-512c-49d0-90db-47de6aaaefd0-system-cni-dir\") pod \"multus-additional-cni-plugins-dx44q\" (UID: \"e5bac22a-512c-49d0-90db-47de6aaaefd0\") " pod="openshift-multus/multus-additional-cni-plugins-dx44q" Dec 06 03:34:23 crc kubenswrapper[4980]: I1206 03:34:23.649040 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/e5bac22a-512c-49d0-90db-47de6aaaefd0-cnibin\") pod \"multus-additional-cni-plugins-dx44q\" (UID: \"e5bac22a-512c-49d0-90db-47de6aaaefd0\") " pod="openshift-multus/multus-additional-cni-plugins-dx44q" Dec 06 03:34:23 crc kubenswrapper[4980]: I1206 03:34:23.649874 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/e5bac22a-512c-49d0-90db-47de6aaaefd0-cni-binary-copy\") pod \"multus-additional-cni-plugins-dx44q\" (UID: \"e5bac22a-512c-49d0-90db-47de6aaaefd0\") " pod="openshift-multus/multus-additional-cni-plugins-dx44q" Dec 06 03:34:23 crc kubenswrapper[4980]: I1206 03:34:23.650170 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/e5bac22a-512c-49d0-90db-47de6aaaefd0-tuning-conf-dir\") pod \"multus-additional-cni-plugins-dx44q\" (UID: \"e5bac22a-512c-49d0-90db-47de6aaaefd0\") " pod="openshift-multus/multus-additional-cni-plugins-dx44q" Dec 06 03:34:23 crc kubenswrapper[4980]: I1206 03:34:23.650598 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/e5bac22a-512c-49d0-90db-47de6aaaefd0-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-dx44q\" (UID: \"e5bac22a-512c-49d0-90db-47de6aaaefd0\") " pod="openshift-multus/multus-additional-cni-plugins-dx44q" Dec 06 03:34:23 crc kubenswrapper[4980]: I1206 03:34:23.667334 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7f958150-549d-4c80-8a7e-23c39e4f9500\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c28b6895b14e73f4f3eb7a8ce93076c05092ba9b58ea0e3e879c65bcfa227a21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://76cff8a2d652df38b04eefc053832baf4cbb5db778ff5e525b18c2825141b118\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://206f6f19fe3197f5d6798cc86c5473bf2decb02a2860ee7ffb5efb5f60952e8c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://773a0bc2aab73ffc2717e987543defc9b786cc22300aa54b4ba4f3454504e331\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://773a0bc2aab73ffc2717e987543defc9b786cc22300aa54b4ba4f3454504e331\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-06T03:34:22Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1206 03:34:15.305330 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1206 03:34:15.307249 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1592846546/tls.crt::/tmp/serving-cert-1592846546/tls.key\\\\\\\"\\\\nI1206 03:34:21.814037 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1206 03:34:21.816756 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1206 03:34:21.816780 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1206 03:34:21.816806 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1206 03:34:21.816813 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1206 03:34:21.831301 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1206 03:34:21.831311 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1206 03:34:21.831324 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 03:34:21.831330 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 03:34:21.831335 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1206 03:34:21.831339 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1206 03:34:21.831342 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1206 03:34:21.831346 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1206 03:34:21.911269 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d40e0e7446f8416bf9da94fc53c01be937f9a51a11eb101157d2ddd4597720d2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:04Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2cba89f2f1bed96d59f183628d900d9b72c91eb472e273c8bed0babcab485910\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2cba89f2f1bed96d59f183628d900d9b72c91eb472e273c8bed0babcab485910\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:01Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 06 03:34:23 crc kubenswrapper[4980]: I1206 03:34:23.674218 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lg7tx\" (UniqueName: \"kubernetes.io/projected/873977aa-4eb0-4526-bc0b-757ad2dea2a5-kube-api-access-lg7tx\") pod \"node-resolver-sv5sz\" (UID: \"873977aa-4eb0-4526-bc0b-757ad2dea2a5\") " pod="openshift-dns/node-resolver-sv5sz" Dec 06 03:34:23 crc kubenswrapper[4980]: I1206 03:34:23.680462 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f7mdm\" (UniqueName: \"kubernetes.io/projected/e5bac22a-512c-49d0-90db-47de6aaaefd0-kube-api-access-f7mdm\") pod \"multus-additional-cni-plugins-dx44q\" (UID: \"e5bac22a-512c-49d0-90db-47de6aaaefd0\") " pod="openshift-multus/multus-additional-cni-plugins-dx44q" Dec 06 03:34:23 crc kubenswrapper[4980]: I1206 03:34:23.689367 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:22Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 06 03:34:23 crc kubenswrapper[4980]: I1206 03:34:23.701189 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:22Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 06 03:34:23 crc kubenswrapper[4980]: I1206 03:34:23.712611 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-sv5sz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"873977aa-4eb0-4526-bc0b-757ad2dea2a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:23Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:23Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lg7tx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:23Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-sv5sz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 06 03:34:23 crc kubenswrapper[4980]: I1206 03:34:23.735327 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5w4fr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5af0e768-3c9c-4401-ab56-57516cd3170e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:23Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:23Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:23Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfzrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfzrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfzrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfzrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfzrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfzrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfzrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfzrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfzrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:23Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-5w4fr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 06 03:34:23 crc kubenswrapper[4980]: I1206 03:34:23.747553 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:22Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 06 03:34:23 crc kubenswrapper[4980]: I1206 03:34:23.749292 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/5af0e768-3c9c-4401-ab56-57516cd3170e-env-overrides\") pod \"ovnkube-node-5w4fr\" (UID: \"5af0e768-3c9c-4401-ab56-57516cd3170e\") " pod="openshift-ovn-kubernetes/ovnkube-node-5w4fr" Dec 06 03:34:23 crc kubenswrapper[4980]: I1206 03:34:23.749345 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/f96c1f6c-55e8-4d86-bff3-5381581b7022-multus-daemon-config\") pod \"multus-kq4z8\" (UID: \"f96c1f6c-55e8-4d86-bff3-5381581b7022\") " pod="openshift-multus/multus-kq4z8" Dec 06 03:34:23 crc kubenswrapper[4980]: I1206 03:34:23.749365 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/5af0e768-3c9c-4401-ab56-57516cd3170e-ovn-node-metrics-cert\") pod \"ovnkube-node-5w4fr\" (UID: \"5af0e768-3c9c-4401-ab56-57516cd3170e\") " pod="openshift-ovn-kubernetes/ovnkube-node-5w4fr" Dec 06 03:34:23 crc kubenswrapper[4980]: I1206 03:34:23.749396 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/f96c1f6c-55e8-4d86-bff3-5381581b7022-multus-cni-dir\") pod \"multus-kq4z8\" (UID: \"f96c1f6c-55e8-4d86-bff3-5381581b7022\") " pod="openshift-multus/multus-kq4z8" Dec 06 03:34:23 crc kubenswrapper[4980]: I1206 03:34:23.749413 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/5af0e768-3c9c-4401-ab56-57516cd3170e-log-socket\") pod \"ovnkube-node-5w4fr\" (UID: \"5af0e768-3c9c-4401-ab56-57516cd3170e\") " pod="openshift-ovn-kubernetes/ovnkube-node-5w4fr" Dec 06 03:34:23 crc kubenswrapper[4980]: I1206 03:34:23.749432 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/5af0e768-3c9c-4401-ab56-57516cd3170e-host-cni-netd\") pod \"ovnkube-node-5w4fr\" (UID: \"5af0e768-3c9c-4401-ab56-57516cd3170e\") " pod="openshift-ovn-kubernetes/ovnkube-node-5w4fr" Dec 06 03:34:23 crc kubenswrapper[4980]: I1206 03:34:23.749457 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/f96c1f6c-55e8-4d86-bff3-5381581b7022-system-cni-dir\") pod \"multus-kq4z8\" (UID: \"f96c1f6c-55e8-4d86-bff3-5381581b7022\") " pod="openshift-multus/multus-kq4z8" Dec 06 03:34:23 crc kubenswrapper[4980]: I1206 03:34:23.749477 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/f96c1f6c-55e8-4d86-bff3-5381581b7022-host-run-netns\") pod \"multus-kq4z8\" (UID: \"f96c1f6c-55e8-4d86-bff3-5381581b7022\") " pod="openshift-multus/multus-kq4z8" Dec 06 03:34:23 crc kubenswrapper[4980]: I1206 03:34:23.749497 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/320f44d1-a671-4a91-b328-a8b0fdd8f23a-proxy-tls\") pod \"machine-config-daemon-r5zfb\" (UID: \"320f44d1-a671-4a91-b328-a8b0fdd8f23a\") " pod="openshift-machine-config-operator/machine-config-daemon-r5zfb" Dec 06 03:34:23 crc kubenswrapper[4980]: I1206 03:34:23.749534 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/f96c1f6c-55e8-4d86-bff3-5381581b7022-host-var-lib-cni-multus\") pod \"multus-kq4z8\" (UID: \"f96c1f6c-55e8-4d86-bff3-5381581b7022\") " pod="openshift-multus/multus-kq4z8" Dec 06 03:34:23 crc kubenswrapper[4980]: I1206 03:34:23.749553 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/5af0e768-3c9c-4401-ab56-57516cd3170e-host-kubelet\") pod \"ovnkube-node-5w4fr\" (UID: \"5af0e768-3c9c-4401-ab56-57516cd3170e\") " pod="openshift-ovn-kubernetes/ovnkube-node-5w4fr" Dec 06 03:34:23 crc kubenswrapper[4980]: I1206 03:34:23.749570 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/5af0e768-3c9c-4401-ab56-57516cd3170e-ovnkube-script-lib\") pod \"ovnkube-node-5w4fr\" (UID: \"5af0e768-3c9c-4401-ab56-57516cd3170e\") " pod="openshift-ovn-kubernetes/ovnkube-node-5w4fr" Dec 06 03:34:23 crc kubenswrapper[4980]: I1206 03:34:23.749588 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/f96c1f6c-55e8-4d86-bff3-5381581b7022-host-var-lib-kubelet\") pod \"multus-kq4z8\" (UID: \"f96c1f6c-55e8-4d86-bff3-5381581b7022\") " pod="openshift-multus/multus-kq4z8" Dec 06 03:34:23 crc kubenswrapper[4980]: I1206 03:34:23.749605 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/5af0e768-3c9c-4401-ab56-57516cd3170e-host-slash\") pod \"ovnkube-node-5w4fr\" (UID: \"5af0e768-3c9c-4401-ab56-57516cd3170e\") " pod="openshift-ovn-kubernetes/ovnkube-node-5w4fr" Dec 06 03:34:23 crc kubenswrapper[4980]: I1206 03:34:23.749628 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/f96c1f6c-55e8-4d86-bff3-5381581b7022-multus-conf-dir\") pod \"multus-kq4z8\" (UID: \"f96c1f6c-55e8-4d86-bff3-5381581b7022\") " pod="openshift-multus/multus-kq4z8" Dec 06 03:34:23 crc kubenswrapper[4980]: I1206 03:34:23.749644 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/5af0e768-3c9c-4401-ab56-57516cd3170e-host-cni-bin\") pod \"ovnkube-node-5w4fr\" (UID: \"5af0e768-3c9c-4401-ab56-57516cd3170e\") " pod="openshift-ovn-kubernetes/ovnkube-node-5w4fr" Dec 06 03:34:23 crc kubenswrapper[4980]: I1206 03:34:23.749660 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/5af0e768-3c9c-4401-ab56-57516cd3170e-node-log\") pod \"ovnkube-node-5w4fr\" (UID: \"5af0e768-3c9c-4401-ab56-57516cd3170e\") " pod="openshift-ovn-kubernetes/ovnkube-node-5w4fr" Dec 06 03:34:23 crc kubenswrapper[4980]: I1206 03:34:23.749679 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/5af0e768-3c9c-4401-ab56-57516cd3170e-run-openvswitch\") pod \"ovnkube-node-5w4fr\" (UID: \"5af0e768-3c9c-4401-ab56-57516cd3170e\") " pod="openshift-ovn-kubernetes/ovnkube-node-5w4fr" Dec 06 03:34:23 crc kubenswrapper[4980]: I1206 03:34:23.749696 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/5af0e768-3c9c-4401-ab56-57516cd3170e-host-run-ovn-kubernetes\") pod \"ovnkube-node-5w4fr\" (UID: \"5af0e768-3c9c-4401-ab56-57516cd3170e\") " pod="openshift-ovn-kubernetes/ovnkube-node-5w4fr" Dec 06 03:34:23 crc kubenswrapper[4980]: I1206 03:34:23.749717 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/f96c1f6c-55e8-4d86-bff3-5381581b7022-cni-binary-copy\") pod \"multus-kq4z8\" (UID: \"f96c1f6c-55e8-4d86-bff3-5381581b7022\") " pod="openshift-multus/multus-kq4z8" Dec 06 03:34:23 crc kubenswrapper[4980]: I1206 03:34:23.749735 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/5af0e768-3c9c-4401-ab56-57516cd3170e-etc-openvswitch\") pod \"ovnkube-node-5w4fr\" (UID: \"5af0e768-3c9c-4401-ab56-57516cd3170e\") " pod="openshift-ovn-kubernetes/ovnkube-node-5w4fr" Dec 06 03:34:23 crc kubenswrapper[4980]: I1206 03:34:23.749758 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/320f44d1-a671-4a91-b328-a8b0fdd8f23a-rootfs\") pod \"machine-config-daemon-r5zfb\" (UID: \"320f44d1-a671-4a91-b328-a8b0fdd8f23a\") " pod="openshift-machine-config-operator/machine-config-daemon-r5zfb" Dec 06 03:34:23 crc kubenswrapper[4980]: I1206 03:34:23.749778 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zhm5t\" (UniqueName: \"kubernetes.io/projected/320f44d1-a671-4a91-b328-a8b0fdd8f23a-kube-api-access-zhm5t\") pod \"machine-config-daemon-r5zfb\" (UID: \"320f44d1-a671-4a91-b328-a8b0fdd8f23a\") " pod="openshift-machine-config-operator/machine-config-daemon-r5zfb" Dec 06 03:34:23 crc kubenswrapper[4980]: I1206 03:34:23.749794 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/5af0e768-3c9c-4401-ab56-57516cd3170e-run-systemd\") pod \"ovnkube-node-5w4fr\" (UID: \"5af0e768-3c9c-4401-ab56-57516cd3170e\") " pod="openshift-ovn-kubernetes/ovnkube-node-5w4fr" Dec 06 03:34:23 crc kubenswrapper[4980]: I1206 03:34:23.749815 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/f96c1f6c-55e8-4d86-bff3-5381581b7022-cnibin\") pod \"multus-kq4z8\" (UID: \"f96c1f6c-55e8-4d86-bff3-5381581b7022\") " pod="openshift-multus/multus-kq4z8" Dec 06 03:34:23 crc kubenswrapper[4980]: I1206 03:34:23.749854 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/f96c1f6c-55e8-4d86-bff3-5381581b7022-host-var-lib-cni-bin\") pod \"multus-kq4z8\" (UID: \"f96c1f6c-55e8-4d86-bff3-5381581b7022\") " pod="openshift-multus/multus-kq4z8" Dec 06 03:34:23 crc kubenswrapper[4980]: I1206 03:34:23.749960 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/f96c1f6c-55e8-4d86-bff3-5381581b7022-etc-kubernetes\") pod \"multus-kq4z8\" (UID: \"f96c1f6c-55e8-4d86-bff3-5381581b7022\") " pod="openshift-multus/multus-kq4z8" Dec 06 03:34:23 crc kubenswrapper[4980]: I1206 03:34:23.749981 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/5af0e768-3c9c-4401-ab56-57516cd3170e-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-5w4fr\" (UID: \"5af0e768-3c9c-4401-ab56-57516cd3170e\") " pod="openshift-ovn-kubernetes/ovnkube-node-5w4fr" Dec 06 03:34:23 crc kubenswrapper[4980]: I1206 03:34:23.750012 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/5af0e768-3c9c-4401-ab56-57516cd3170e-ovnkube-config\") pod \"ovnkube-node-5w4fr\" (UID: \"5af0e768-3c9c-4401-ab56-57516cd3170e\") " pod="openshift-ovn-kubernetes/ovnkube-node-5w4fr" Dec 06 03:34:23 crc kubenswrapper[4980]: I1206 03:34:23.750039 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/f96c1f6c-55e8-4d86-bff3-5381581b7022-hostroot\") pod \"multus-kq4z8\" (UID: \"f96c1f6c-55e8-4d86-bff3-5381581b7022\") " pod="openshift-multus/multus-kq4z8" Dec 06 03:34:23 crc kubenswrapper[4980]: I1206 03:34:23.750061 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/320f44d1-a671-4a91-b328-a8b0fdd8f23a-mcd-auth-proxy-config\") pod \"machine-config-daemon-r5zfb\" (UID: \"320f44d1-a671-4a91-b328-a8b0fdd8f23a\") " pod="openshift-machine-config-operator/machine-config-daemon-r5zfb" Dec 06 03:34:23 crc kubenswrapper[4980]: I1206 03:34:23.750079 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/5af0e768-3c9c-4401-ab56-57516cd3170e-var-lib-openvswitch\") pod \"ovnkube-node-5w4fr\" (UID: \"5af0e768-3c9c-4401-ab56-57516cd3170e\") " pod="openshift-ovn-kubernetes/ovnkube-node-5w4fr" Dec 06 03:34:23 crc kubenswrapper[4980]: I1206 03:34:23.750096 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/f96c1f6c-55e8-4d86-bff3-5381581b7022-host-run-multus-certs\") pod \"multus-kq4z8\" (UID: \"f96c1f6c-55e8-4d86-bff3-5381581b7022\") " pod="openshift-multus/multus-kq4z8" Dec 06 03:34:23 crc kubenswrapper[4980]: I1206 03:34:23.750117 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bdx7g\" (UniqueName: \"kubernetes.io/projected/f96c1f6c-55e8-4d86-bff3-5381581b7022-kube-api-access-bdx7g\") pod \"multus-kq4z8\" (UID: \"f96c1f6c-55e8-4d86-bff3-5381581b7022\") " pod="openshift-multus/multus-kq4z8" Dec 06 03:34:23 crc kubenswrapper[4980]: I1206 03:34:23.750134 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/5af0e768-3c9c-4401-ab56-57516cd3170e-run-ovn\") pod \"ovnkube-node-5w4fr\" (UID: \"5af0e768-3c9c-4401-ab56-57516cd3170e\") " pod="openshift-ovn-kubernetes/ovnkube-node-5w4fr" Dec 06 03:34:23 crc kubenswrapper[4980]: I1206 03:34:23.750151 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/f96c1f6c-55e8-4d86-bff3-5381581b7022-os-release\") pod \"multus-kq4z8\" (UID: \"f96c1f6c-55e8-4d86-bff3-5381581b7022\") " pod="openshift-multus/multus-kq4z8" Dec 06 03:34:23 crc kubenswrapper[4980]: I1206 03:34:23.750175 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/f96c1f6c-55e8-4d86-bff3-5381581b7022-multus-socket-dir-parent\") pod \"multus-kq4z8\" (UID: \"f96c1f6c-55e8-4d86-bff3-5381581b7022\") " pod="openshift-multus/multus-kq4z8" Dec 06 03:34:23 crc kubenswrapper[4980]: I1206 03:34:23.750212 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/f96c1f6c-55e8-4d86-bff3-5381581b7022-host-run-k8s-cni-cncf-io\") pod \"multus-kq4z8\" (UID: \"f96c1f6c-55e8-4d86-bff3-5381581b7022\") " pod="openshift-multus/multus-kq4z8" Dec 06 03:34:23 crc kubenswrapper[4980]: I1206 03:34:23.750240 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zfzrb\" (UniqueName: \"kubernetes.io/projected/5af0e768-3c9c-4401-ab56-57516cd3170e-kube-api-access-zfzrb\") pod \"ovnkube-node-5w4fr\" (UID: \"5af0e768-3c9c-4401-ab56-57516cd3170e\") " pod="openshift-ovn-kubernetes/ovnkube-node-5w4fr" Dec 06 03:34:23 crc kubenswrapper[4980]: I1206 03:34:23.750276 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/5af0e768-3c9c-4401-ab56-57516cd3170e-systemd-units\") pod \"ovnkube-node-5w4fr\" (UID: \"5af0e768-3c9c-4401-ab56-57516cd3170e\") " pod="openshift-ovn-kubernetes/ovnkube-node-5w4fr" Dec 06 03:34:23 crc kubenswrapper[4980]: I1206 03:34:23.750285 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/5af0e768-3c9c-4401-ab56-57516cd3170e-env-overrides\") pod \"ovnkube-node-5w4fr\" (UID: \"5af0e768-3c9c-4401-ab56-57516cd3170e\") " pod="openshift-ovn-kubernetes/ovnkube-node-5w4fr" Dec 06 03:34:23 crc kubenswrapper[4980]: I1206 03:34:23.750294 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/5af0e768-3c9c-4401-ab56-57516cd3170e-host-run-netns\") pod \"ovnkube-node-5w4fr\" (UID: \"5af0e768-3c9c-4401-ab56-57516cd3170e\") " pod="openshift-ovn-kubernetes/ovnkube-node-5w4fr" Dec 06 03:34:23 crc kubenswrapper[4980]: I1206 03:34:23.750358 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/5af0e768-3c9c-4401-ab56-57516cd3170e-host-run-netns\") pod \"ovnkube-node-5w4fr\" (UID: \"5af0e768-3c9c-4401-ab56-57516cd3170e\") " pod="openshift-ovn-kubernetes/ovnkube-node-5w4fr" Dec 06 03:34:23 crc kubenswrapper[4980]: I1206 03:34:23.750551 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/5af0e768-3c9c-4401-ab56-57516cd3170e-log-socket\") pod \"ovnkube-node-5w4fr\" (UID: \"5af0e768-3c9c-4401-ab56-57516cd3170e\") " pod="openshift-ovn-kubernetes/ovnkube-node-5w4fr" Dec 06 03:34:23 crc kubenswrapper[4980]: I1206 03:34:23.750599 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/5af0e768-3c9c-4401-ab56-57516cd3170e-host-cni-netd\") pod \"ovnkube-node-5w4fr\" (UID: \"5af0e768-3c9c-4401-ab56-57516cd3170e\") " pod="openshift-ovn-kubernetes/ovnkube-node-5w4fr" Dec 06 03:34:23 crc kubenswrapper[4980]: I1206 03:34:23.750795 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/f96c1f6c-55e8-4d86-bff3-5381581b7022-multus-cni-dir\") pod \"multus-kq4z8\" (UID: \"f96c1f6c-55e8-4d86-bff3-5381581b7022\") " pod="openshift-multus/multus-kq4z8" Dec 06 03:34:23 crc kubenswrapper[4980]: I1206 03:34:23.750859 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/5af0e768-3c9c-4401-ab56-57516cd3170e-run-systemd\") pod \"ovnkube-node-5w4fr\" (UID: \"5af0e768-3c9c-4401-ab56-57516cd3170e\") " pod="openshift-ovn-kubernetes/ovnkube-node-5w4fr" Dec 06 03:34:23 crc kubenswrapper[4980]: I1206 03:34:23.750978 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/f96c1f6c-55e8-4d86-bff3-5381581b7022-host-run-netns\") pod \"multus-kq4z8\" (UID: \"f96c1f6c-55e8-4d86-bff3-5381581b7022\") " pod="openshift-multus/multus-kq4z8" Dec 06 03:34:23 crc kubenswrapper[4980]: I1206 03:34:23.751065 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/f96c1f6c-55e8-4d86-bff3-5381581b7022-system-cni-dir\") pod \"multus-kq4z8\" (UID: \"f96c1f6c-55e8-4d86-bff3-5381581b7022\") " pod="openshift-multus/multus-kq4z8" Dec 06 03:34:23 crc kubenswrapper[4980]: I1206 03:34:23.751193 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/f96c1f6c-55e8-4d86-bff3-5381581b7022-cnibin\") pod \"multus-kq4z8\" (UID: \"f96c1f6c-55e8-4d86-bff3-5381581b7022\") " pod="openshift-multus/multus-kq4z8" Dec 06 03:34:23 crc kubenswrapper[4980]: I1206 03:34:23.751265 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/f96c1f6c-55e8-4d86-bff3-5381581b7022-multus-socket-dir-parent\") pod \"multus-kq4z8\" (UID: \"f96c1f6c-55e8-4d86-bff3-5381581b7022\") " pod="openshift-multus/multus-kq4z8" Dec 06 03:34:23 crc kubenswrapper[4980]: I1206 03:34:23.751284 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/f96c1f6c-55e8-4d86-bff3-5381581b7022-multus-daemon-config\") pod \"multus-kq4z8\" (UID: \"f96c1f6c-55e8-4d86-bff3-5381581b7022\") " pod="openshift-multus/multus-kq4z8" Dec 06 03:34:23 crc kubenswrapper[4980]: I1206 03:34:23.751298 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/f96c1f6c-55e8-4d86-bff3-5381581b7022-host-run-multus-certs\") pod \"multus-kq4z8\" (UID: \"f96c1f6c-55e8-4d86-bff3-5381581b7022\") " pod="openshift-multus/multus-kq4z8" Dec 06 03:34:23 crc kubenswrapper[4980]: I1206 03:34:23.751424 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/f96c1f6c-55e8-4d86-bff3-5381581b7022-host-var-lib-cni-bin\") pod \"multus-kq4z8\" (UID: \"f96c1f6c-55e8-4d86-bff3-5381581b7022\") " pod="openshift-multus/multus-kq4z8" Dec 06 03:34:23 crc kubenswrapper[4980]: I1206 03:34:23.751482 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/f96c1f6c-55e8-4d86-bff3-5381581b7022-host-run-k8s-cni-cncf-io\") pod \"multus-kq4z8\" (UID: \"f96c1f6c-55e8-4d86-bff3-5381581b7022\") " pod="openshift-multus/multus-kq4z8" Dec 06 03:34:23 crc kubenswrapper[4980]: I1206 03:34:23.751481 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/5af0e768-3c9c-4401-ab56-57516cd3170e-host-cni-bin\") pod \"ovnkube-node-5w4fr\" (UID: \"5af0e768-3c9c-4401-ab56-57516cd3170e\") " pod="openshift-ovn-kubernetes/ovnkube-node-5w4fr" Dec 06 03:34:23 crc kubenswrapper[4980]: I1206 03:34:23.751539 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/f96c1f6c-55e8-4d86-bff3-5381581b7022-host-var-lib-cni-multus\") pod \"multus-kq4z8\" (UID: \"f96c1f6c-55e8-4d86-bff3-5381581b7022\") " pod="openshift-multus/multus-kq4z8" Dec 06 03:34:23 crc kubenswrapper[4980]: I1206 03:34:23.751605 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/f96c1f6c-55e8-4d86-bff3-5381581b7022-os-release\") pod \"multus-kq4z8\" (UID: \"f96c1f6c-55e8-4d86-bff3-5381581b7022\") " pod="openshift-multus/multus-kq4z8" Dec 06 03:34:23 crc kubenswrapper[4980]: I1206 03:34:23.751575 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/5af0e768-3c9c-4401-ab56-57516cd3170e-host-kubelet\") pod \"ovnkube-node-5w4fr\" (UID: \"5af0e768-3c9c-4401-ab56-57516cd3170e\") " pod="openshift-ovn-kubernetes/ovnkube-node-5w4fr" Dec 06 03:34:23 crc kubenswrapper[4980]: I1206 03:34:23.752231 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/5af0e768-3c9c-4401-ab56-57516cd3170e-ovnkube-config\") pod \"ovnkube-node-5w4fr\" (UID: \"5af0e768-3c9c-4401-ab56-57516cd3170e\") " pod="openshift-ovn-kubernetes/ovnkube-node-5w4fr" Dec 06 03:34:23 crc kubenswrapper[4980]: I1206 03:34:23.752277 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/f96c1f6c-55e8-4d86-bff3-5381581b7022-etc-kubernetes\") pod \"multus-kq4z8\" (UID: \"f96c1f6c-55e8-4d86-bff3-5381581b7022\") " pod="openshift-multus/multus-kq4z8" Dec 06 03:34:23 crc kubenswrapper[4980]: I1206 03:34:23.752304 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/5af0e768-3c9c-4401-ab56-57516cd3170e-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-5w4fr\" (UID: \"5af0e768-3c9c-4401-ab56-57516cd3170e\") " pod="openshift-ovn-kubernetes/ovnkube-node-5w4fr" Dec 06 03:34:23 crc kubenswrapper[4980]: I1206 03:34:23.752334 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/5af0e768-3c9c-4401-ab56-57516cd3170e-host-slash\") pod \"ovnkube-node-5w4fr\" (UID: \"5af0e768-3c9c-4401-ab56-57516cd3170e\") " pod="openshift-ovn-kubernetes/ovnkube-node-5w4fr" Dec 06 03:34:23 crc kubenswrapper[4980]: I1206 03:34:23.752357 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/f96c1f6c-55e8-4d86-bff3-5381581b7022-host-var-lib-kubelet\") pod \"multus-kq4z8\" (UID: \"f96c1f6c-55e8-4d86-bff3-5381581b7022\") " pod="openshift-multus/multus-kq4z8" Dec 06 03:34:23 crc kubenswrapper[4980]: I1206 03:34:23.752415 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/f96c1f6c-55e8-4d86-bff3-5381581b7022-cni-binary-copy\") pod \"multus-kq4z8\" (UID: \"f96c1f6c-55e8-4d86-bff3-5381581b7022\") " pod="openshift-multus/multus-kq4z8" Dec 06 03:34:23 crc kubenswrapper[4980]: I1206 03:34:23.752416 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/5af0e768-3c9c-4401-ab56-57516cd3170e-ovnkube-script-lib\") pod \"ovnkube-node-5w4fr\" (UID: \"5af0e768-3c9c-4401-ab56-57516cd3170e\") " pod="openshift-ovn-kubernetes/ovnkube-node-5w4fr" Dec 06 03:34:23 crc kubenswrapper[4980]: I1206 03:34:23.752442 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/f96c1f6c-55e8-4d86-bff3-5381581b7022-multus-conf-dir\") pod \"multus-kq4z8\" (UID: \"f96c1f6c-55e8-4d86-bff3-5381581b7022\") " pod="openshift-multus/multus-kq4z8" Dec 06 03:34:23 crc kubenswrapper[4980]: I1206 03:34:23.752479 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/320f44d1-a671-4a91-b328-a8b0fdd8f23a-rootfs\") pod \"machine-config-daemon-r5zfb\" (UID: \"320f44d1-a671-4a91-b328-a8b0fdd8f23a\") " pod="openshift-machine-config-operator/machine-config-daemon-r5zfb" Dec 06 03:34:23 crc kubenswrapper[4980]: I1206 03:34:23.752488 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/5af0e768-3c9c-4401-ab56-57516cd3170e-node-log\") pod \"ovnkube-node-5w4fr\" (UID: \"5af0e768-3c9c-4401-ab56-57516cd3170e\") " pod="openshift-ovn-kubernetes/ovnkube-node-5w4fr" Dec 06 03:34:23 crc kubenswrapper[4980]: I1206 03:34:23.752525 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/5af0e768-3c9c-4401-ab56-57516cd3170e-etc-openvswitch\") pod \"ovnkube-node-5w4fr\" (UID: \"5af0e768-3c9c-4401-ab56-57516cd3170e\") " pod="openshift-ovn-kubernetes/ovnkube-node-5w4fr" Dec 06 03:34:23 crc kubenswrapper[4980]: I1206 03:34:23.752529 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/5af0e768-3c9c-4401-ab56-57516cd3170e-run-openvswitch\") pod \"ovnkube-node-5w4fr\" (UID: \"5af0e768-3c9c-4401-ab56-57516cd3170e\") " pod="openshift-ovn-kubernetes/ovnkube-node-5w4fr" Dec 06 03:34:23 crc kubenswrapper[4980]: I1206 03:34:23.752553 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/5af0e768-3c9c-4401-ab56-57516cd3170e-var-lib-openvswitch\") pod \"ovnkube-node-5w4fr\" (UID: \"5af0e768-3c9c-4401-ab56-57516cd3170e\") " pod="openshift-ovn-kubernetes/ovnkube-node-5w4fr" Dec 06 03:34:23 crc kubenswrapper[4980]: I1206 03:34:23.752558 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/5af0e768-3c9c-4401-ab56-57516cd3170e-host-run-ovn-kubernetes\") pod \"ovnkube-node-5w4fr\" (UID: \"5af0e768-3c9c-4401-ab56-57516cd3170e\") " pod="openshift-ovn-kubernetes/ovnkube-node-5w4fr" Dec 06 03:34:23 crc kubenswrapper[4980]: I1206 03:34:23.752582 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/5af0e768-3c9c-4401-ab56-57516cd3170e-systemd-units\") pod \"ovnkube-node-5w4fr\" (UID: \"5af0e768-3c9c-4401-ab56-57516cd3170e\") " pod="openshift-ovn-kubernetes/ovnkube-node-5w4fr" Dec 06 03:34:23 crc kubenswrapper[4980]: I1206 03:34:23.752589 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/f96c1f6c-55e8-4d86-bff3-5381581b7022-hostroot\") pod \"multus-kq4z8\" (UID: \"f96c1f6c-55e8-4d86-bff3-5381581b7022\") " pod="openshift-multus/multus-kq4z8" Dec 06 03:34:23 crc kubenswrapper[4980]: I1206 03:34:23.751536 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/5af0e768-3c9c-4401-ab56-57516cd3170e-run-ovn\") pod \"ovnkube-node-5w4fr\" (UID: \"5af0e768-3c9c-4401-ab56-57516cd3170e\") " pod="openshift-ovn-kubernetes/ovnkube-node-5w4fr" Dec 06 03:34:23 crc kubenswrapper[4980]: I1206 03:34:23.752829 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/320f44d1-a671-4a91-b328-a8b0fdd8f23a-mcd-auth-proxy-config\") pod \"machine-config-daemon-r5zfb\" (UID: \"320f44d1-a671-4a91-b328-a8b0fdd8f23a\") " pod="openshift-machine-config-operator/machine-config-daemon-r5zfb" Dec 06 03:34:23 crc kubenswrapper[4980]: I1206 03:34:23.753161 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/5af0e768-3c9c-4401-ab56-57516cd3170e-ovn-node-metrics-cert\") pod \"ovnkube-node-5w4fr\" (UID: \"5af0e768-3c9c-4401-ab56-57516cd3170e\") " pod="openshift-ovn-kubernetes/ovnkube-node-5w4fr" Dec 06 03:34:23 crc kubenswrapper[4980]: I1206 03:34:23.755282 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/320f44d1-a671-4a91-b328-a8b0fdd8f23a-proxy-tls\") pod \"machine-config-daemon-r5zfb\" (UID: \"320f44d1-a671-4a91-b328-a8b0fdd8f23a\") " pod="openshift-machine-config-operator/machine-config-daemon-r5zfb" Dec 06 03:34:23 crc kubenswrapper[4980]: I1206 03:34:23.763294 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:22Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 06 03:34:23 crc kubenswrapper[4980]: I1206 03:34:23.822196 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bdx7g\" (UniqueName: \"kubernetes.io/projected/f96c1f6c-55e8-4d86-bff3-5381581b7022-kube-api-access-bdx7g\") pod \"multus-kq4z8\" (UID: \"f96c1f6c-55e8-4d86-bff3-5381581b7022\") " pod="openshift-multus/multus-kq4z8" Dec 06 03:34:23 crc kubenswrapper[4980]: I1206 03:34:23.824495 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zfzrb\" (UniqueName: \"kubernetes.io/projected/5af0e768-3c9c-4401-ab56-57516cd3170e-kube-api-access-zfzrb\") pod \"ovnkube-node-5w4fr\" (UID: \"5af0e768-3c9c-4401-ab56-57516cd3170e\") " pod="openshift-ovn-kubernetes/ovnkube-node-5w4fr" Dec 06 03:34:23 crc kubenswrapper[4980]: I1206 03:34:23.828798 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-r5zfb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"320f44d1-a671-4a91-b328-a8b0fdd8f23a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:23Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:23Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zhm5t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zhm5t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:23Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-r5zfb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 06 03:34:23 crc kubenswrapper[4980]: I1206 03:34:23.829461 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zhm5t\" (UniqueName: \"kubernetes.io/projected/320f44d1-a671-4a91-b328-a8b0fdd8f23a-kube-api-access-zhm5t\") pod \"machine-config-daemon-r5zfb\" (UID: \"320f44d1-a671-4a91-b328-a8b0fdd8f23a\") " pod="openshift-machine-config-operator/machine-config-daemon-r5zfb" Dec 06 03:34:23 crc kubenswrapper[4980]: I1206 03:34:23.851621 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 03:34:23 crc kubenswrapper[4980]: I1206 03:34:23.851689 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 03:34:23 crc kubenswrapper[4980]: E1206 03:34:23.851842 4980 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 06 03:34:23 crc kubenswrapper[4980]: E1206 03:34:23.852106 4980 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-06 03:34:25.852029343 +0000 UTC m=+25.096780614 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 06 03:34:23 crc kubenswrapper[4980]: E1206 03:34:23.852175 4980 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 06 03:34:23 crc kubenswrapper[4980]: E1206 03:34:23.852318 4980 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-06 03:34:25.85229619 +0000 UTC m=+25.097047461 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 06 03:34:23 crc kubenswrapper[4980]: I1206 03:34:23.852433 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cab608b3-a02f-475b-9aae-9b8f5d878de9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f7fcc44b42bcbb55509048e3a566bb557333c4a83e33d7b0fba8eb514317909\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c22e7dd8886c2dfbb742900ed86ec7b0c2edf9a1363e74f6666961d928f06c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://473e792b2d0a9644854ee0581c6120fc9c989bc342899ef72b1d056080fd4694\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ee8e936dfc39361d6f99460910591ca8c901a20f0d274bbb6aaebfa8a00ee06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://77bc12768b018625ba11a6bfd66688f926c4b7f3bc1d0e41c6551a18b485ecf3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://27cf07cd96893900da4afecd0d56fd9269bc60c1d23246b4685fcd8756357c19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://27cf07cd96893900da4afecd0d56fd9269bc60c1d23246b4685fcd8756357c19\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c49feb24566a8284378670fa0390710d38c237acc94ffb654f90f3fa4626e4c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c49feb24566a8284378670fa0390710d38c237acc94ffb654f90f3fa4626e4c8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:03Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://bb74d1c3d752a379fa394c450217badb45511fbcafad62f7b618131cab94b29e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bb74d1c3d752a379fa394c450217badb45511fbcafad62f7b618131cab94b29e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:01Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 06 03:34:23 crc kubenswrapper[4980]: I1206 03:34:23.869414 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:22Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 06 03:34:23 crc kubenswrapper[4980]: I1206 03:34:23.880528 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-dx44q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5bac22a-512c-49d0-90db-47de6aaaefd0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:23Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:23Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:23Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f7mdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f7mdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f7mdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f7mdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f7mdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f7mdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f7mdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:23Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-dx44q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 06 03:34:23 crc kubenswrapper[4980]: I1206 03:34:23.889025 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-sv5sz" Dec 06 03:34:23 crc kubenswrapper[4980]: I1206 03:34:23.896043 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 06 03:34:23 crc kubenswrapper[4980]: W1206 03:34:23.903731 4980 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod873977aa_4eb0_4526_bc0b_757ad2dea2a5.slice/crio-a29a15088c34aa009b88f7b93aace1ecfda3342b043fea96de05e47f630b9788 WatchSource:0}: Error finding container a29a15088c34aa009b88f7b93aace1ecfda3342b043fea96de05e47f630b9788: Status 404 returned error can't find the container with id a29a15088c34aa009b88f7b93aace1ecfda3342b043fea96de05e47f630b9788 Dec 06 03:34:23 crc kubenswrapper[4980]: I1206 03:34:23.904655 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-dx44q" Dec 06 03:34:23 crc kubenswrapper[4980]: I1206 03:34:23.909263 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 06 03:34:23 crc kubenswrapper[4980]: I1206 03:34:23.925635 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-kq4z8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f96c1f6c-55e8-4d86-bff3-5381581b7022\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:23Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:23Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bdx7g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:23Z\\\"}}\" for pod \"openshift-multus\"/\"multus-kq4z8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 06 03:34:23 crc kubenswrapper[4980]: I1206 03:34:23.926608 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-kq4z8" Dec 06 03:34:23 crc kubenswrapper[4980]: I1206 03:34:23.947040 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-5w4fr" Dec 06 03:34:23 crc kubenswrapper[4980]: I1206 03:34:23.947616 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-r5zfb" Dec 06 03:34:23 crc kubenswrapper[4980]: I1206 03:34:23.953225 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 03:34:23 crc kubenswrapper[4980]: I1206 03:34:23.953362 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 03:34:23 crc kubenswrapper[4980]: I1206 03:34:23.953420 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 03:34:23 crc kubenswrapper[4980]: E1206 03:34:23.953600 4980 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 06 03:34:23 crc kubenswrapper[4980]: E1206 03:34:23.953624 4980 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 06 03:34:23 crc kubenswrapper[4980]: E1206 03:34:23.953638 4980 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 06 03:34:23 crc kubenswrapper[4980]: E1206 03:34:23.953696 4980 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-06 03:34:25.953679312 +0000 UTC m=+25.198430583 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 06 03:34:23 crc kubenswrapper[4980]: E1206 03:34:23.953758 4980 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 03:34:25.953751294 +0000 UTC m=+25.198502565 (durationBeforeRetry 2s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 03:34:23 crc kubenswrapper[4980]: E1206 03:34:23.953826 4980 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 06 03:34:23 crc kubenswrapper[4980]: E1206 03:34:23.953837 4980 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 06 03:34:23 crc kubenswrapper[4980]: E1206 03:34:23.953846 4980 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 06 03:34:23 crc kubenswrapper[4980]: E1206 03:34:23.953871 4980 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-06 03:34:25.953863367 +0000 UTC m=+25.198614638 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 06 03:34:24 crc kubenswrapper[4980]: I1206 03:34:24.017227 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7f958150-549d-4c80-8a7e-23c39e4f9500\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c28b6895b14e73f4f3eb7a8ce93076c05092ba9b58ea0e3e879c65bcfa227a21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://76cff8a2d652df38b04eefc053832baf4cbb5db778ff5e525b18c2825141b118\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://206f6f19fe3197f5d6798cc86c5473bf2decb02a2860ee7ffb5efb5f60952e8c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bafaa9670b05671fb57b658c2960bf3ee32c84e52f86a3a3bbab01d5a38ff0c4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://773a0bc2aab73ffc2717e987543defc9b786cc22300aa54b4ba4f3454504e331\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-06T03:34:22Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1206 03:34:15.305330 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1206 03:34:15.307249 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1592846546/tls.crt::/tmp/serving-cert-1592846546/tls.key\\\\\\\"\\\\nI1206 03:34:21.814037 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1206 03:34:21.816756 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1206 03:34:21.816780 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1206 03:34:21.816806 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1206 03:34:21.816813 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1206 03:34:21.831301 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1206 03:34:21.831311 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1206 03:34:21.831324 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 03:34:21.831330 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 03:34:21.831335 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1206 03:34:21.831339 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1206 03:34:21.831342 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1206 03:34:21.831346 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1206 03:34:21.911269 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:04Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d40e0e7446f8416bf9da94fc53c01be937f9a51a11eb101157d2ddd4597720d2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:04Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2cba89f2f1bed96d59f183628d900d9b72c91eb472e273c8bed0babcab485910\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2cba89f2f1bed96d59f183628d900d9b72c91eb472e273c8bed0babcab485910\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:01Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 06 03:34:24 crc kubenswrapper[4980]: W1206 03:34:24.060095 4980 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5af0e768_3c9c_4401_ab56_57516cd3170e.slice/crio-c6155f097866dc8f452b08fe7fb793ca7d6cf707ecfbb096967ad970e78f6cb9 WatchSource:0}: Error finding container c6155f097866dc8f452b08fe7fb793ca7d6cf707ecfbb096967ad970e78f6cb9: Status 404 returned error can't find the container with id c6155f097866dc8f452b08fe7fb793ca7d6cf707ecfbb096967ad970e78f6cb9 Dec 06 03:34:24 crc kubenswrapper[4980]: I1206 03:34:24.183698 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 03:34:24 crc kubenswrapper[4980]: E1206 03:34:24.183831 4980 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 03:34:24 crc kubenswrapper[4980]: I1206 03:34:24.183913 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 03:34:24 crc kubenswrapper[4980]: E1206 03:34:24.183966 4980 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 03:34:24 crc kubenswrapper[4980]: I1206 03:34:24.527503 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-kq4z8" event={"ID":"f96c1f6c-55e8-4d86-bff3-5381581b7022","Type":"ContainerStarted","Data":"799498681afa7653408bc23b1e82d75c16e3c809d2ee7f322780e20c537eb710"} Dec 06 03:34:24 crc kubenswrapper[4980]: I1206 03:34:24.527599 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-kq4z8" event={"ID":"f96c1f6c-55e8-4d86-bff3-5381581b7022","Type":"ContainerStarted","Data":"60fbb678f06d62aa91d0a22d9bed65f283bd4b70965993fde433f4ef60f2bc7d"} Dec 06 03:34:24 crc kubenswrapper[4980]: I1206 03:34:24.530340 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-sv5sz" event={"ID":"873977aa-4eb0-4526-bc0b-757ad2dea2a5","Type":"ContainerStarted","Data":"01f0cc855c822a71ea8ef6a7ce6e25747ff9956691735973f935c4a28672a7a6"} Dec 06 03:34:24 crc kubenswrapper[4980]: I1206 03:34:24.530410 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-sv5sz" event={"ID":"873977aa-4eb0-4526-bc0b-757ad2dea2a5","Type":"ContainerStarted","Data":"a29a15088c34aa009b88f7b93aace1ecfda3342b043fea96de05e47f630b9788"} Dec 06 03:34:24 crc kubenswrapper[4980]: I1206 03:34:24.531825 4980 generic.go:334] "Generic (PLEG): container finished" podID="5af0e768-3c9c-4401-ab56-57516cd3170e" containerID="d876df04b5ab0ba07a5cbcafbe4648403c5014c59eb7d40adee2657c3a255473" exitCode=0 Dec 06 03:34:24 crc kubenswrapper[4980]: I1206 03:34:24.531896 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5w4fr" event={"ID":"5af0e768-3c9c-4401-ab56-57516cd3170e","Type":"ContainerDied","Data":"d876df04b5ab0ba07a5cbcafbe4648403c5014c59eb7d40adee2657c3a255473"} Dec 06 03:34:24 crc kubenswrapper[4980]: I1206 03:34:24.531930 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5w4fr" event={"ID":"5af0e768-3c9c-4401-ab56-57516cd3170e","Type":"ContainerStarted","Data":"c6155f097866dc8f452b08fe7fb793ca7d6cf707ecfbb096967ad970e78f6cb9"} Dec 06 03:34:24 crc kubenswrapper[4980]: I1206 03:34:24.533573 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"e4ef01e7d9809fd9cfc46e3a58b07159cdc6f5a3c210d1c9efc49bca27b30858"} Dec 06 03:34:24 crc kubenswrapper[4980]: I1206 03:34:24.535581 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-r5zfb" event={"ID":"320f44d1-a671-4a91-b328-a8b0fdd8f23a","Type":"ContainerStarted","Data":"92393ecaaef8cfcfd01837671f8cc89803dd570b637b4a373b40b7f487788826"} Dec 06 03:34:24 crc kubenswrapper[4980]: I1206 03:34:24.535641 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-r5zfb" event={"ID":"320f44d1-a671-4a91-b328-a8b0fdd8f23a","Type":"ContainerStarted","Data":"52e8db644463794041978ced2c0591c13e0ada8e5f1a78a94a0a665218a37b5f"} Dec 06 03:34:24 crc kubenswrapper[4980]: I1206 03:34:24.535655 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-r5zfb" event={"ID":"320f44d1-a671-4a91-b328-a8b0fdd8f23a","Type":"ContainerStarted","Data":"e0cbd719a28685991dbaa33ccaadfc02618e4a0a36fdddb003e1299a93393661"} Dec 06 03:34:24 crc kubenswrapper[4980]: I1206 03:34:24.537442 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-dx44q" event={"ID":"e5bac22a-512c-49d0-90db-47de6aaaefd0","Type":"ContainerStarted","Data":"a4142f8059503f97386a8af97e23329550c4393813c69b69681b112bc23fda8b"} Dec 06 03:34:24 crc kubenswrapper[4980]: I1206 03:34:24.607242 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:24Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:24 crc kubenswrapper[4980]: I1206 03:34:24.783123 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:24Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:24 crc kubenswrapper[4980]: I1206 03:34:24.850647 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-kq4z8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f96c1f6c-55e8-4d86-bff3-5381581b7022\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://799498681afa7653408bc23b1e82d75c16e3c809d2ee7f322780e20c537eb710\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bdx7g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:23Z\\\"}}\" for pod \"openshift-multus\"/\"multus-kq4z8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:24Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:24 crc kubenswrapper[4980]: I1206 03:34:24.888110 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7f958150-549d-4c80-8a7e-23c39e4f9500\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c28b6895b14e73f4f3eb7a8ce93076c05092ba9b58ea0e3e879c65bcfa227a21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://76cff8a2d652df38b04eefc053832baf4cbb5db778ff5e525b18c2825141b118\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://206f6f19fe3197f5d6798cc86c5473bf2decb02a2860ee7ffb5efb5f60952e8c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bafaa9670b05671fb57b658c2960bf3ee32c84e52f86a3a3bbab01d5a38ff0c4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://773a0bc2aab73ffc2717e987543defc9b786cc22300aa54b4ba4f3454504e331\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-06T03:34:22Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1206 03:34:15.305330 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1206 03:34:15.307249 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1592846546/tls.crt::/tmp/serving-cert-1592846546/tls.key\\\\\\\"\\\\nI1206 03:34:21.814037 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1206 03:34:21.816756 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1206 03:34:21.816780 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1206 03:34:21.816806 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1206 03:34:21.816813 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1206 03:34:21.831301 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1206 03:34:21.831311 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1206 03:34:21.831324 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 03:34:21.831330 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 03:34:21.831335 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1206 03:34:21.831339 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1206 03:34:21.831342 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1206 03:34:21.831346 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1206 03:34:21.911269 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:04Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d40e0e7446f8416bf9da94fc53c01be937f9a51a11eb101157d2ddd4597720d2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:04Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2cba89f2f1bed96d59f183628d900d9b72c91eb472e273c8bed0babcab485910\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2cba89f2f1bed96d59f183628d900d9b72c91eb472e273c8bed0babcab485910\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:01Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:24Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:24 crc kubenswrapper[4980]: I1206 03:34:24.917881 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:22Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:24Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:24 crc kubenswrapper[4980]: I1206 03:34:24.929924 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-sv5sz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"873977aa-4eb0-4526-bc0b-757ad2dea2a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:23Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:23Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lg7tx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:23Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-sv5sz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:24Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:24 crc kubenswrapper[4980]: I1206 03:34:24.956846 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5w4fr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5af0e768-3c9c-4401-ab56-57516cd3170e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:23Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:23Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:23Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfzrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfzrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfzrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfzrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfzrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfzrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfzrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfzrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfzrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:23Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-5w4fr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:24Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:24 crc kubenswrapper[4980]: I1206 03:34:24.971185 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:22Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:24Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:24 crc kubenswrapper[4980]: I1206 03:34:24.984777 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:22Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:24Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:25 crc kubenswrapper[4980]: I1206 03:34:25.001083 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-r5zfb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"320f44d1-a671-4a91-b328-a8b0fdd8f23a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:23Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:23Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zhm5t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zhm5t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:23Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-r5zfb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:24Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:25 crc kubenswrapper[4980]: I1206 03:34:25.023182 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cab608b3-a02f-475b-9aae-9b8f5d878de9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f7fcc44b42bcbb55509048e3a566bb557333c4a83e33d7b0fba8eb514317909\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c22e7dd8886c2dfbb742900ed86ec7b0c2edf9a1363e74f6666961d928f06c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://473e792b2d0a9644854ee0581c6120fc9c989bc342899ef72b1d056080fd4694\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ee8e936dfc39361d6f99460910591ca8c901a20f0d274bbb6aaebfa8a00ee06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://77bc12768b018625ba11a6bfd66688f926c4b7f3bc1d0e41c6551a18b485ecf3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://27cf07cd96893900da4afecd0d56fd9269bc60c1d23246b4685fcd8756357c19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://27cf07cd96893900da4afecd0d56fd9269bc60c1d23246b4685fcd8756357c19\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c49feb24566a8284378670fa0390710d38c237acc94ffb654f90f3fa4626e4c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c49feb24566a8284378670fa0390710d38c237acc94ffb654f90f3fa4626e4c8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:03Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://bb74d1c3d752a379fa394c450217badb45511fbcafad62f7b618131cab94b29e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bb74d1c3d752a379fa394c450217badb45511fbcafad62f7b618131cab94b29e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:01Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:25Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:25 crc kubenswrapper[4980]: I1206 03:34:25.034939 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:22Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:25Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:25 crc kubenswrapper[4980]: I1206 03:34:25.050159 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-dx44q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5bac22a-512c-49d0-90db-47de6aaaefd0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:23Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:23Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:23Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f7mdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f7mdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f7mdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f7mdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f7mdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f7mdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f7mdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:23Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-dx44q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:25Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:25 crc kubenswrapper[4980]: I1206 03:34:25.062373 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:22Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:25Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:25 crc kubenswrapper[4980]: I1206 03:34:25.077473 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://334ea6a273fd7d268906c9f4a6dbf6f4df4fcdbf38f3d7d248cb7d6cf8537794\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ec44dcd5e4bf0a0466bb586ebd61259364dbbfe22499f4cbff854e3bd153082\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:25Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:25 crc kubenswrapper[4980]: I1206 03:34:25.092277 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:22Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:25Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:25 crc kubenswrapper[4980]: I1206 03:34:25.103082 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-sv5sz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"873977aa-4eb0-4526-bc0b-757ad2dea2a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01f0cc855c822a71ea8ef6a7ce6e25747ff9956691735973f935c4a28672a7a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lg7tx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:23Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-sv5sz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:25Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:25 crc kubenswrapper[4980]: I1206 03:34:25.119315 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5w4fr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5af0e768-3c9c-4401-ab56-57516cd3170e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:23Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:23Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfzrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfzrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfzrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfzrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfzrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfzrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfzrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfzrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d876df04b5ab0ba07a5cbcafbe4648403c5014c59eb7d40adee2657c3a255473\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d876df04b5ab0ba07a5cbcafbe4648403c5014c59eb7d40adee2657c3a255473\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfzrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:23Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-5w4fr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:25Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:25 crc kubenswrapper[4980]: I1206 03:34:25.130859 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:22Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:25Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:25 crc kubenswrapper[4980]: I1206 03:34:25.148799 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-dx44q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5bac22a-512c-49d0-90db-47de6aaaefd0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:23Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:23Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:23Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f7mdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f7mdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f7mdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f7mdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f7mdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f7mdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f7mdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:23Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-dx44q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:25Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:25 crc kubenswrapper[4980]: I1206 03:34:25.167538 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-r5zfb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"320f44d1-a671-4a91-b328-a8b0fdd8f23a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://92393ecaaef8cfcfd01837671f8cc89803dd570b637b4a373b40b7f487788826\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zhm5t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52e8db644463794041978ced2c0591c13e0ada8e5f1a78a94a0a665218a37b5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zhm5t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:23Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-r5zfb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:25Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:25 crc kubenswrapper[4980]: I1206 03:34:25.193351 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 03:34:25 crc kubenswrapper[4980]: E1206 03:34:25.193474 4980 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 03:34:25 crc kubenswrapper[4980]: I1206 03:34:25.194229 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cab608b3-a02f-475b-9aae-9b8f5d878de9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f7fcc44b42bcbb55509048e3a566bb557333c4a83e33d7b0fba8eb514317909\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c22e7dd8886c2dfbb742900ed86ec7b0c2edf9a1363e74f6666961d928f06c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://473e792b2d0a9644854ee0581c6120fc9c989bc342899ef72b1d056080fd4694\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ee8e936dfc39361d6f99460910591ca8c901a20f0d274bbb6aaebfa8a00ee06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://77bc12768b018625ba11a6bfd66688f926c4b7f3bc1d0e41c6551a18b485ecf3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://27cf07cd96893900da4afecd0d56fd9269bc60c1d23246b4685fcd8756357c19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://27cf07cd96893900da4afecd0d56fd9269bc60c1d23246b4685fcd8756357c19\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c49feb24566a8284378670fa0390710d38c237acc94ffb654f90f3fa4626e4c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c49feb24566a8284378670fa0390710d38c237acc94ffb654f90f3fa4626e4c8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:03Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://bb74d1c3d752a379fa394c450217badb45511fbcafad62f7b618131cab94b29e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bb74d1c3d752a379fa394c450217badb45511fbcafad62f7b618131cab94b29e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:01Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:25Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:25 crc kubenswrapper[4980]: I1206 03:34:25.208781 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4ef01e7d9809fd9cfc46e3a58b07159cdc6f5a3c210d1c9efc49bca27b30858\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:25Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:25 crc kubenswrapper[4980]: I1206 03:34:25.301616 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-kq4z8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f96c1f6c-55e8-4d86-bff3-5381581b7022\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://799498681afa7653408bc23b1e82d75c16e3c809d2ee7f322780e20c537eb710\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bdx7g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:23Z\\\"}}\" for pod \"openshift-multus\"/\"multus-kq4z8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:25Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:25 crc kubenswrapper[4980]: I1206 03:34:25.316134 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:25Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:25 crc kubenswrapper[4980]: I1206 03:34:25.330143 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7f958150-549d-4c80-8a7e-23c39e4f9500\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c28b6895b14e73f4f3eb7a8ce93076c05092ba9b58ea0e3e879c65bcfa227a21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://76cff8a2d652df38b04eefc053832baf4cbb5db778ff5e525b18c2825141b118\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://206f6f19fe3197f5d6798cc86c5473bf2decb02a2860ee7ffb5efb5f60952e8c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bafaa9670b05671fb57b658c2960bf3ee32c84e52f86a3a3bbab01d5a38ff0c4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://773a0bc2aab73ffc2717e987543defc9b786cc22300aa54b4ba4f3454504e331\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-06T03:34:22Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1206 03:34:15.305330 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1206 03:34:15.307249 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1592846546/tls.crt::/tmp/serving-cert-1592846546/tls.key\\\\\\\"\\\\nI1206 03:34:21.814037 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1206 03:34:21.816756 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1206 03:34:21.816780 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1206 03:34:21.816806 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1206 03:34:21.816813 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1206 03:34:21.831301 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1206 03:34:21.831311 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1206 03:34:21.831324 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 03:34:21.831330 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 03:34:21.831335 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1206 03:34:21.831339 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1206 03:34:21.831342 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1206 03:34:21.831346 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1206 03:34:21.911269 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:04Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d40e0e7446f8416bf9da94fc53c01be937f9a51a11eb101157d2ddd4597720d2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:04Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2cba89f2f1bed96d59f183628d900d9b72c91eb472e273c8bed0babcab485910\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2cba89f2f1bed96d59f183628d900d9b72c91eb472e273c8bed0babcab485910\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:01Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:25Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:25 crc kubenswrapper[4980]: I1206 03:34:25.547331 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5w4fr" event={"ID":"5af0e768-3c9c-4401-ab56-57516cd3170e","Type":"ContainerStarted","Data":"768b8e1fe1a00a8467d744da5ad3f61006b0a3612c96f8e4d0ee6fcb93195f9f"} Dec 06 03:34:25 crc kubenswrapper[4980]: I1206 03:34:25.547386 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5w4fr" event={"ID":"5af0e768-3c9c-4401-ab56-57516cd3170e","Type":"ContainerStarted","Data":"f35beae5339eeaf26494ed678e3c1bd02349eae2a4b4615325b17923f3b653b9"} Dec 06 03:34:25 crc kubenswrapper[4980]: I1206 03:34:25.547402 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5w4fr" event={"ID":"5af0e768-3c9c-4401-ab56-57516cd3170e","Type":"ContainerStarted","Data":"a940ec6ee523313728f7e8ae934b7e0b0cff82771e3e64723309db1fae013c3a"} Dec 06 03:34:25 crc kubenswrapper[4980]: I1206 03:34:25.549417 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"502c727d6baa31f400bee14c34b19abbe38af873f28f6a82a43547431755d981"} Dec 06 03:34:25 crc kubenswrapper[4980]: I1206 03:34:25.551079 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-dx44q" event={"ID":"e5bac22a-512c-49d0-90db-47de6aaaefd0","Type":"ContainerStarted","Data":"c3b5dab476c7f227a6e4a2eae8469432667fb9fad345d9419ba59feab9fd385e"} Dec 06 03:34:25 crc kubenswrapper[4980]: I1206 03:34:25.571617 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cab608b3-a02f-475b-9aae-9b8f5d878de9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f7fcc44b42bcbb55509048e3a566bb557333c4a83e33d7b0fba8eb514317909\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c22e7dd8886c2dfbb742900ed86ec7b0c2edf9a1363e74f6666961d928f06c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://473e792b2d0a9644854ee0581c6120fc9c989bc342899ef72b1d056080fd4694\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ee8e936dfc39361d6f99460910591ca8c901a20f0d274bbb6aaebfa8a00ee06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://77bc12768b018625ba11a6bfd66688f926c4b7f3bc1d0e41c6551a18b485ecf3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://27cf07cd96893900da4afecd0d56fd9269bc60c1d23246b4685fcd8756357c19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://27cf07cd96893900da4afecd0d56fd9269bc60c1d23246b4685fcd8756357c19\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c49feb24566a8284378670fa0390710d38c237acc94ffb654f90f3fa4626e4c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c49feb24566a8284378670fa0390710d38c237acc94ffb654f90f3fa4626e4c8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:03Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://bb74d1c3d752a379fa394c450217badb45511fbcafad62f7b618131cab94b29e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bb74d1c3d752a379fa394c450217badb45511fbcafad62f7b618131cab94b29e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:01Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:25Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:25 crc kubenswrapper[4980]: I1206 03:34:25.587008 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://502c727d6baa31f400bee14c34b19abbe38af873f28f6a82a43547431755d981\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:25Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:25 crc kubenswrapper[4980]: I1206 03:34:25.604141 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-dx44q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5bac22a-512c-49d0-90db-47de6aaaefd0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:23Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:23Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:23Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f7mdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f7mdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f7mdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f7mdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f7mdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f7mdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f7mdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:23Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-dx44q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:25Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:25 crc kubenswrapper[4980]: I1206 03:34:25.616727 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-r5zfb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"320f44d1-a671-4a91-b328-a8b0fdd8f23a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://92393ecaaef8cfcfd01837671f8cc89803dd570b637b4a373b40b7f487788826\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zhm5t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52e8db644463794041978ced2c0591c13e0ada8e5f1a78a94a0a665218a37b5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zhm5t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:23Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-r5zfb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:25Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:25 crc kubenswrapper[4980]: I1206 03:34:25.639759 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:25Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:25 crc kubenswrapper[4980]: I1206 03:34:25.658956 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4ef01e7d9809fd9cfc46e3a58b07159cdc6f5a3c210d1c9efc49bca27b30858\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:25Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:25 crc kubenswrapper[4980]: I1206 03:34:25.680239 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-kq4z8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f96c1f6c-55e8-4d86-bff3-5381581b7022\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://799498681afa7653408bc23b1e82d75c16e3c809d2ee7f322780e20c537eb710\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bdx7g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:23Z\\\"}}\" for pod \"openshift-multus\"/\"multus-kq4z8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:25Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:25 crc kubenswrapper[4980]: I1206 03:34:25.714243 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7f958150-549d-4c80-8a7e-23c39e4f9500\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c28b6895b14e73f4f3eb7a8ce93076c05092ba9b58ea0e3e879c65bcfa227a21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://76cff8a2d652df38b04eefc053832baf4cbb5db778ff5e525b18c2825141b118\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://206f6f19fe3197f5d6798cc86c5473bf2decb02a2860ee7ffb5efb5f60952e8c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bafaa9670b05671fb57b658c2960bf3ee32c84e52f86a3a3bbab01d5a38ff0c4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://773a0bc2aab73ffc2717e987543defc9b786cc22300aa54b4ba4f3454504e331\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-06T03:34:22Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1206 03:34:15.305330 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1206 03:34:15.307249 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1592846546/tls.crt::/tmp/serving-cert-1592846546/tls.key\\\\\\\"\\\\nI1206 03:34:21.814037 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1206 03:34:21.816756 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1206 03:34:21.816780 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1206 03:34:21.816806 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1206 03:34:21.816813 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1206 03:34:21.831301 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1206 03:34:21.831311 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1206 03:34:21.831324 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 03:34:21.831330 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 03:34:21.831335 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1206 03:34:21.831339 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1206 03:34:21.831342 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1206 03:34:21.831346 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1206 03:34:21.911269 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:04Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d40e0e7446f8416bf9da94fc53c01be937f9a51a11eb101157d2ddd4597720d2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:04Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2cba89f2f1bed96d59f183628d900d9b72c91eb472e273c8bed0babcab485910\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2cba89f2f1bed96d59f183628d900d9b72c91eb472e273c8bed0babcab485910\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:01Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:25Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:25 crc kubenswrapper[4980]: I1206 03:34:25.733792 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5w4fr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5af0e768-3c9c-4401-ab56-57516cd3170e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:23Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:23Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfzrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfzrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfzrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfzrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfzrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfzrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfzrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfzrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d876df04b5ab0ba07a5cbcafbe4648403c5014c59eb7d40adee2657c3a255473\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d876df04b5ab0ba07a5cbcafbe4648403c5014c59eb7d40adee2657c3a255473\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfzrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:23Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-5w4fr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:25Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:25 crc kubenswrapper[4980]: I1206 03:34:25.768418 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:22Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:25Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:25 crc kubenswrapper[4980]: I1206 03:34:25.782429 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://334ea6a273fd7d268906c9f4a6dbf6f4df4fcdbf38f3d7d248cb7d6cf8537794\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ec44dcd5e4bf0a0466bb586ebd61259364dbbfe22499f4cbff854e3bd153082\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:25Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:25 crc kubenswrapper[4980]: I1206 03:34:25.844601 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:22Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:25Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:25 crc kubenswrapper[4980]: I1206 03:34:25.858339 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-sv5sz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"873977aa-4eb0-4526-bc0b-757ad2dea2a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01f0cc855c822a71ea8ef6a7ce6e25747ff9956691735973f935c4a28672a7a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lg7tx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:23Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-sv5sz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:25Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:25 crc kubenswrapper[4980]: I1206 03:34:25.869314 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:25Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:25 crc kubenswrapper[4980]: I1206 03:34:25.880296 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4ef01e7d9809fd9cfc46e3a58b07159cdc6f5a3c210d1c9efc49bca27b30858\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:25Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:25 crc kubenswrapper[4980]: I1206 03:34:25.891507 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-kq4z8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f96c1f6c-55e8-4d86-bff3-5381581b7022\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://799498681afa7653408bc23b1e82d75c16e3c809d2ee7f322780e20c537eb710\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bdx7g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:23Z\\\"}}\" for pod \"openshift-multus\"/\"multus-kq4z8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:25Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:25 crc kubenswrapper[4980]: I1206 03:34:25.903611 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7f958150-549d-4c80-8a7e-23c39e4f9500\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c28b6895b14e73f4f3eb7a8ce93076c05092ba9b58ea0e3e879c65bcfa227a21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://76cff8a2d652df38b04eefc053832baf4cbb5db778ff5e525b18c2825141b118\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://206f6f19fe3197f5d6798cc86c5473bf2decb02a2860ee7ffb5efb5f60952e8c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bafaa9670b05671fb57b658c2960bf3ee32c84e52f86a3a3bbab01d5a38ff0c4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://773a0bc2aab73ffc2717e987543defc9b786cc22300aa54b4ba4f3454504e331\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-06T03:34:22Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1206 03:34:15.305330 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1206 03:34:15.307249 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1592846546/tls.crt::/tmp/serving-cert-1592846546/tls.key\\\\\\\"\\\\nI1206 03:34:21.814037 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1206 03:34:21.816756 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1206 03:34:21.816780 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1206 03:34:21.816806 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1206 03:34:21.816813 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1206 03:34:21.831301 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1206 03:34:21.831311 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1206 03:34:21.831324 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 03:34:21.831330 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 03:34:21.831335 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1206 03:34:21.831339 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1206 03:34:21.831342 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1206 03:34:21.831346 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1206 03:34:21.911269 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:04Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d40e0e7446f8416bf9da94fc53c01be937f9a51a11eb101157d2ddd4597720d2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:04Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2cba89f2f1bed96d59f183628d900d9b72c91eb472e273c8bed0babcab485910\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2cba89f2f1bed96d59f183628d900d9b72c91eb472e273c8bed0babcab485910\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:01Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:25Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:25 crc kubenswrapper[4980]: I1206 03:34:25.904862 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 03:34:25 crc kubenswrapper[4980]: I1206 03:34:25.904910 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 03:34:25 crc kubenswrapper[4980]: E1206 03:34:25.905035 4980 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 06 03:34:25 crc kubenswrapper[4980]: E1206 03:34:25.905061 4980 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 06 03:34:25 crc kubenswrapper[4980]: E1206 03:34:25.905118 4980 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-06 03:34:29.905091294 +0000 UTC m=+29.149842565 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 06 03:34:25 crc kubenswrapper[4980]: E1206 03:34:25.905161 4980 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-06 03:34:29.905140605 +0000 UTC m=+29.149891876 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 06 03:34:25 crc kubenswrapper[4980]: I1206 03:34:25.922408 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:22Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:25Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:25 crc kubenswrapper[4980]: I1206 03:34:25.935053 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-sv5sz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"873977aa-4eb0-4526-bc0b-757ad2dea2a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01f0cc855c822a71ea8ef6a7ce6e25747ff9956691735973f935c4a28672a7a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lg7tx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:23Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-sv5sz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:25Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:25 crc kubenswrapper[4980]: I1206 03:34:25.952851 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5w4fr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5af0e768-3c9c-4401-ab56-57516cd3170e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:23Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:23Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfzrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfzrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfzrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfzrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfzrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfzrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfzrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfzrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d876df04b5ab0ba07a5cbcafbe4648403c5014c59eb7d40adee2657c3a255473\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d876df04b5ab0ba07a5cbcafbe4648403c5014c59eb7d40adee2657c3a255473\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfzrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:23Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-5w4fr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:25Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:25 crc kubenswrapper[4980]: I1206 03:34:25.966776 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:22Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:25Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:25 crc kubenswrapper[4980]: I1206 03:34:25.983696 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://334ea6a273fd7d268906c9f4a6dbf6f4df4fcdbf38f3d7d248cb7d6cf8537794\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ec44dcd5e4bf0a0466bb586ebd61259364dbbfe22499f4cbff854e3bd153082\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:25Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:26 crc kubenswrapper[4980]: I1206 03:34:26.002598 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-r5zfb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"320f44d1-a671-4a91-b328-a8b0fdd8f23a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://92393ecaaef8cfcfd01837671f8cc89803dd570b637b4a373b40b7f487788826\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zhm5t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52e8db644463794041978ced2c0591c13e0ada8e5f1a78a94a0a665218a37b5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zhm5t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:23Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-r5zfb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:25Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:26 crc kubenswrapper[4980]: I1206 03:34:26.005955 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 03:34:26 crc kubenswrapper[4980]: I1206 03:34:26.006040 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 03:34:26 crc kubenswrapper[4980]: I1206 03:34:26.006073 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 03:34:26 crc kubenswrapper[4980]: E1206 03:34:26.006110 4980 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 03:34:30.006094406 +0000 UTC m=+29.250845677 (durationBeforeRetry 4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 03:34:26 crc kubenswrapper[4980]: E1206 03:34:26.006184 4980 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 06 03:34:26 crc kubenswrapper[4980]: E1206 03:34:26.006198 4980 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 06 03:34:26 crc kubenswrapper[4980]: E1206 03:34:26.006208 4980 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 06 03:34:26 crc kubenswrapper[4980]: E1206 03:34:26.006237 4980 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-06 03:34:30.00622977 +0000 UTC m=+29.250981041 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 06 03:34:26 crc kubenswrapper[4980]: E1206 03:34:26.006283 4980 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 06 03:34:26 crc kubenswrapper[4980]: E1206 03:34:26.006332 4980 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 06 03:34:26 crc kubenswrapper[4980]: E1206 03:34:26.006346 4980 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 06 03:34:26 crc kubenswrapper[4980]: E1206 03:34:26.006428 4980 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-06 03:34:30.006401885 +0000 UTC m=+29.251153156 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 06 03:34:26 crc kubenswrapper[4980]: I1206 03:34:26.026606 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cab608b3-a02f-475b-9aae-9b8f5d878de9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f7fcc44b42bcbb55509048e3a566bb557333c4a83e33d7b0fba8eb514317909\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c22e7dd8886c2dfbb742900ed86ec7b0c2edf9a1363e74f6666961d928f06c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://473e792b2d0a9644854ee0581c6120fc9c989bc342899ef72b1d056080fd4694\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ee8e936dfc39361d6f99460910591ca8c901a20f0d274bbb6aaebfa8a00ee06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://77bc12768b018625ba11a6bfd66688f926c4b7f3bc1d0e41c6551a18b485ecf3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://27cf07cd96893900da4afecd0d56fd9269bc60c1d23246b4685fcd8756357c19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://27cf07cd96893900da4afecd0d56fd9269bc60c1d23246b4685fcd8756357c19\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c49feb24566a8284378670fa0390710d38c237acc94ffb654f90f3fa4626e4c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c49feb24566a8284378670fa0390710d38c237acc94ffb654f90f3fa4626e4c8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:03Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://bb74d1c3d752a379fa394c450217badb45511fbcafad62f7b618131cab94b29e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bb74d1c3d752a379fa394c450217badb45511fbcafad62f7b618131cab94b29e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:01Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:26Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:26 crc kubenswrapper[4980]: I1206 03:34:26.041526 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://502c727d6baa31f400bee14c34b19abbe38af873f28f6a82a43547431755d981\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:26Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:26 crc kubenswrapper[4980]: I1206 03:34:26.056405 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-dx44q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5bac22a-512c-49d0-90db-47de6aaaefd0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:23Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:23Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:23Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f7mdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3b5dab476c7f227a6e4a2eae8469432667fb9fad345d9419ba59feab9fd385e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f7mdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f7mdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f7mdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f7mdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f7mdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f7mdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:23Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-dx44q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:26Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:26 crc kubenswrapper[4980]: I1206 03:34:26.217549 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 03:34:26 crc kubenswrapper[4980]: I1206 03:34:26.217630 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 03:34:26 crc kubenswrapper[4980]: E1206 03:34:26.217996 4980 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 03:34:26 crc kubenswrapper[4980]: E1206 03:34:26.218003 4980 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 03:34:26 crc kubenswrapper[4980]: I1206 03:34:26.558133 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5w4fr" event={"ID":"5af0e768-3c9c-4401-ab56-57516cd3170e","Type":"ContainerStarted","Data":"6342903b318317ee8d84c2f5d467cb25c811a789be0d6ab0e0cba37b5611e16d"} Dec 06 03:34:27 crc kubenswrapper[4980]: I1206 03:34:27.186006 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 03:34:27 crc kubenswrapper[4980]: E1206 03:34:27.186378 4980 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 03:34:27 crc kubenswrapper[4980]: I1206 03:34:27.584652 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5w4fr" event={"ID":"5af0e768-3c9c-4401-ab56-57516cd3170e","Type":"ContainerStarted","Data":"2d620685574db7594132ae24a91ec0a93dce20a9a289540deebb39650b8a1816"} Dec 06 03:34:27 crc kubenswrapper[4980]: I1206 03:34:27.584700 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5w4fr" event={"ID":"5af0e768-3c9c-4401-ab56-57516cd3170e","Type":"ContainerStarted","Data":"c67ec8f9c1764fcdf215d52822c2d4f0437dc7f056f05d708767121f6c1f3515"} Dec 06 03:34:28 crc kubenswrapper[4980]: I1206 03:34:28.183698 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 03:34:28 crc kubenswrapper[4980]: I1206 03:34:28.183806 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 03:34:28 crc kubenswrapper[4980]: E1206 03:34:28.183829 4980 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 03:34:28 crc kubenswrapper[4980]: E1206 03:34:28.183976 4980 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 03:34:28 crc kubenswrapper[4980]: I1206 03:34:28.209344 4980 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 03:34:28 crc kubenswrapper[4980]: I1206 03:34:28.246915 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:34:28 crc kubenswrapper[4980]: I1206 03:34:28.246950 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:34:28 crc kubenswrapper[4980]: I1206 03:34:28.246959 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:34:28 crc kubenswrapper[4980]: I1206 03:34:28.247047 4980 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 06 03:34:28 crc kubenswrapper[4980]: I1206 03:34:28.255581 4980 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 06 03:34:28 crc kubenswrapper[4980]: I1206 03:34:28.256939 4980 kubelet_node_status.go:115] "Node was previously registered" node="crc" Dec 06 03:34:28 crc kubenswrapper[4980]: I1206 03:34:28.257239 4980 kubelet_node_status.go:79] "Successfully registered node" node="crc" Dec 06 03:34:28 crc kubenswrapper[4980]: I1206 03:34:28.258363 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:34:28 crc kubenswrapper[4980]: I1206 03:34:28.258391 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:34:28 crc kubenswrapper[4980]: I1206 03:34:28.258400 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:34:28 crc kubenswrapper[4980]: I1206 03:34:28.258412 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:34:28 crc kubenswrapper[4980]: I1206 03:34:28.258422 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:34:28Z","lastTransitionTime":"2025-12-06T03:34:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:34:28 crc kubenswrapper[4980]: I1206 03:34:28.260601 4980 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 06 03:34:28 crc kubenswrapper[4980]: I1206 03:34:28.268382 4980 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/kube-controller-manager-crc"] Dec 06 03:34:28 crc kubenswrapper[4980]: I1206 03:34:28.278621 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:28Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:28 crc kubenswrapper[4980]: E1206 03:34:28.290086 4980 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T03:34:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T03:34:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:28Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T03:34:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T03:34:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:28Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"9b4ed4f8-ed8d-447b-9e1f-13defbdcf0f7\\\",\\\"systemUUID\\\":\\\"a5b493aa-69de-4688-8ff6-1e5f49b7f014\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:28Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:28 crc kubenswrapper[4980]: I1206 03:34:28.294264 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:34:28 crc kubenswrapper[4980]: I1206 03:34:28.294308 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:34:28 crc kubenswrapper[4980]: I1206 03:34:28.294320 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:34:28 crc kubenswrapper[4980]: I1206 03:34:28.294336 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:34:28 crc kubenswrapper[4980]: I1206 03:34:28.294348 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:34:28Z","lastTransitionTime":"2025-12-06T03:34:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:34:28 crc kubenswrapper[4980]: I1206 03:34:28.296400 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4ef01e7d9809fd9cfc46e3a58b07159cdc6f5a3c210d1c9efc49bca27b30858\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:28Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:28 crc kubenswrapper[4980]: E1206 03:34:28.310474 4980 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T03:34:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T03:34:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:28Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T03:34:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T03:34:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:28Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"9b4ed4f8-ed8d-447b-9e1f-13defbdcf0f7\\\",\\\"systemUUID\\\":\\\"a5b493aa-69de-4688-8ff6-1e5f49b7f014\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:28Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:28 crc kubenswrapper[4980]: I1206 03:34:28.313788 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-kq4z8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f96c1f6c-55e8-4d86-bff3-5381581b7022\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://799498681afa7653408bc23b1e82d75c16e3c809d2ee7f322780e20c537eb710\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bdx7g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:23Z\\\"}}\" for pod \"openshift-multus\"/\"multus-kq4z8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:28Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:28 crc kubenswrapper[4980]: I1206 03:34:28.315402 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:34:28 crc kubenswrapper[4980]: I1206 03:34:28.315448 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:34:28 crc kubenswrapper[4980]: I1206 03:34:28.315461 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:34:28 crc kubenswrapper[4980]: I1206 03:34:28.315477 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:34:28 crc kubenswrapper[4980]: I1206 03:34:28.315486 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:34:28Z","lastTransitionTime":"2025-12-06T03:34:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:34:28 crc kubenswrapper[4980]: E1206 03:34:28.329757 4980 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T03:34:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T03:34:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:28Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T03:34:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T03:34:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:28Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"9b4ed4f8-ed8d-447b-9e1f-13defbdcf0f7\\\",\\\"systemUUID\\\":\\\"a5b493aa-69de-4688-8ff6-1e5f49b7f014\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:28Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:28 crc kubenswrapper[4980]: I1206 03:34:28.332554 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7f958150-549d-4c80-8a7e-23c39e4f9500\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c28b6895b14e73f4f3eb7a8ce93076c05092ba9b58ea0e3e879c65bcfa227a21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://76cff8a2d652df38b04eefc053832baf4cbb5db778ff5e525b18c2825141b118\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://206f6f19fe3197f5d6798cc86c5473bf2decb02a2860ee7ffb5efb5f60952e8c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bafaa9670b05671fb57b658c2960bf3ee32c84e52f86a3a3bbab01d5a38ff0c4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://773a0bc2aab73ffc2717e987543defc9b786cc22300aa54b4ba4f3454504e331\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-06T03:34:22Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1206 03:34:15.305330 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1206 03:34:15.307249 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1592846546/tls.crt::/tmp/serving-cert-1592846546/tls.key\\\\\\\"\\\\nI1206 03:34:21.814037 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1206 03:34:21.816756 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1206 03:34:21.816780 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1206 03:34:21.816806 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1206 03:34:21.816813 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1206 03:34:21.831301 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1206 03:34:21.831311 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1206 03:34:21.831324 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 03:34:21.831330 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 03:34:21.831335 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1206 03:34:21.831339 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1206 03:34:21.831342 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1206 03:34:21.831346 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1206 03:34:21.911269 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:04Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d40e0e7446f8416bf9da94fc53c01be937f9a51a11eb101157d2ddd4597720d2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:04Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2cba89f2f1bed96d59f183628d900d9b72c91eb472e273c8bed0babcab485910\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2cba89f2f1bed96d59f183628d900d9b72c91eb472e273c8bed0babcab485910\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:01Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:28Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:28 crc kubenswrapper[4980]: I1206 03:34:28.333292 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:34:28 crc kubenswrapper[4980]: I1206 03:34:28.333344 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:34:28 crc kubenswrapper[4980]: I1206 03:34:28.333359 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:34:28 crc kubenswrapper[4980]: I1206 03:34:28.333379 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:34:28 crc kubenswrapper[4980]: I1206 03:34:28.333391 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:34:28Z","lastTransitionTime":"2025-12-06T03:34:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:34:28 crc kubenswrapper[4980]: I1206 03:34:28.346882 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:22Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:28Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:28 crc kubenswrapper[4980]: E1206 03:34:28.346953 4980 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T03:34:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T03:34:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:28Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T03:34:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T03:34:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:28Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"9b4ed4f8-ed8d-447b-9e1f-13defbdcf0f7\\\",\\\"systemUUID\\\":\\\"a5b493aa-69de-4688-8ff6-1e5f49b7f014\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:28Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:28 crc kubenswrapper[4980]: I1206 03:34:28.351009 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:34:28 crc kubenswrapper[4980]: I1206 03:34:28.351060 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:34:28 crc kubenswrapper[4980]: I1206 03:34:28.351074 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:34:28 crc kubenswrapper[4980]: I1206 03:34:28.351092 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:34:28 crc kubenswrapper[4980]: I1206 03:34:28.351104 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:34:28Z","lastTransitionTime":"2025-12-06T03:34:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:34:28 crc kubenswrapper[4980]: I1206 03:34:28.358197 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-sv5sz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"873977aa-4eb0-4526-bc0b-757ad2dea2a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01f0cc855c822a71ea8ef6a7ce6e25747ff9956691735973f935c4a28672a7a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lg7tx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:23Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-sv5sz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:28Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:28 crc kubenswrapper[4980]: E1206 03:34:28.362977 4980 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T03:34:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T03:34:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:28Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T03:34:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T03:34:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:28Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"9b4ed4f8-ed8d-447b-9e1f-13defbdcf0f7\\\",\\\"systemUUID\\\":\\\"a5b493aa-69de-4688-8ff6-1e5f49b7f014\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:28Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:28 crc kubenswrapper[4980]: E1206 03:34:28.363130 4980 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 06 03:34:28 crc kubenswrapper[4980]: I1206 03:34:28.365141 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:34:28 crc kubenswrapper[4980]: I1206 03:34:28.365183 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:34:28 crc kubenswrapper[4980]: I1206 03:34:28.365200 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:34:28 crc kubenswrapper[4980]: I1206 03:34:28.365217 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:34:28 crc kubenswrapper[4980]: I1206 03:34:28.365228 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:34:28Z","lastTransitionTime":"2025-12-06T03:34:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:34:28 crc kubenswrapper[4980]: I1206 03:34:28.375926 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5w4fr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5af0e768-3c9c-4401-ab56-57516cd3170e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:23Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:23Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfzrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfzrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfzrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfzrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfzrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfzrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfzrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfzrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d876df04b5ab0ba07a5cbcafbe4648403c5014c59eb7d40adee2657c3a255473\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d876df04b5ab0ba07a5cbcafbe4648403c5014c59eb7d40adee2657c3a255473\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfzrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:23Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-5w4fr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:28Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:28 crc kubenswrapper[4980]: I1206 03:34:28.388853 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:22Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:28Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:28 crc kubenswrapper[4980]: I1206 03:34:28.402692 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://334ea6a273fd7d268906c9f4a6dbf6f4df4fcdbf38f3d7d248cb7d6cf8537794\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ec44dcd5e4bf0a0466bb586ebd61259364dbbfe22499f4cbff854e3bd153082\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:28Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:28 crc kubenswrapper[4980]: I1206 03:34:28.416494 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-r5zfb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"320f44d1-a671-4a91-b328-a8b0fdd8f23a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://92393ecaaef8cfcfd01837671f8cc89803dd570b637b4a373b40b7f487788826\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zhm5t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52e8db644463794041978ced2c0591c13e0ada8e5f1a78a94a0a665218a37b5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zhm5t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:23Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-r5zfb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:28Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:28 crc kubenswrapper[4980]: I1206 03:34:28.433298 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cab608b3-a02f-475b-9aae-9b8f5d878de9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f7fcc44b42bcbb55509048e3a566bb557333c4a83e33d7b0fba8eb514317909\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c22e7dd8886c2dfbb742900ed86ec7b0c2edf9a1363e74f6666961d928f06c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://473e792b2d0a9644854ee0581c6120fc9c989bc342899ef72b1d056080fd4694\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ee8e936dfc39361d6f99460910591ca8c901a20f0d274bbb6aaebfa8a00ee06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://77bc12768b018625ba11a6bfd66688f926c4b7f3bc1d0e41c6551a18b485ecf3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://27cf07cd96893900da4afecd0d56fd9269bc60c1d23246b4685fcd8756357c19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://27cf07cd96893900da4afecd0d56fd9269bc60c1d23246b4685fcd8756357c19\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c49feb24566a8284378670fa0390710d38c237acc94ffb654f90f3fa4626e4c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c49feb24566a8284378670fa0390710d38c237acc94ffb654f90f3fa4626e4c8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:03Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://bb74d1c3d752a379fa394c450217badb45511fbcafad62f7b618131cab94b29e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bb74d1c3d752a379fa394c450217badb45511fbcafad62f7b618131cab94b29e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:01Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:28Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:28 crc kubenswrapper[4980]: I1206 03:34:28.445003 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://502c727d6baa31f400bee14c34b19abbe38af873f28f6a82a43547431755d981\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:28Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:28 crc kubenswrapper[4980]: I1206 03:34:28.458629 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-dx44q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5bac22a-512c-49d0-90db-47de6aaaefd0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:23Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:23Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:23Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f7mdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3b5dab476c7f227a6e4a2eae8469432667fb9fad345d9419ba59feab9fd385e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f7mdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f7mdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f7mdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f7mdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f7mdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f7mdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:23Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-dx44q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:28Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:28 crc kubenswrapper[4980]: I1206 03:34:28.471284 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:34:28 crc kubenswrapper[4980]: I1206 03:34:28.471332 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:34:28 crc kubenswrapper[4980]: I1206 03:34:28.471344 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:34:28 crc kubenswrapper[4980]: I1206 03:34:28.471364 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:34:28 crc kubenswrapper[4980]: I1206 03:34:28.471376 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:34:28Z","lastTransitionTime":"2025-12-06T03:34:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:34:28 crc kubenswrapper[4980]: I1206 03:34:28.475225 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:28Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:28 crc kubenswrapper[4980]: I1206 03:34:28.487828 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4ef01e7d9809fd9cfc46e3a58b07159cdc6f5a3c210d1c9efc49bca27b30858\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:28Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:28 crc kubenswrapper[4980]: I1206 03:34:28.500399 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-kq4z8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f96c1f6c-55e8-4d86-bff3-5381581b7022\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://799498681afa7653408bc23b1e82d75c16e3c809d2ee7f322780e20c537eb710\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bdx7g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:23Z\\\"}}\" for pod \"openshift-multus\"/\"multus-kq4z8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:28Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:28 crc kubenswrapper[4980]: I1206 03:34:28.516140 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7f958150-549d-4c80-8a7e-23c39e4f9500\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c28b6895b14e73f4f3eb7a8ce93076c05092ba9b58ea0e3e879c65bcfa227a21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://76cff8a2d652df38b04eefc053832baf4cbb5db778ff5e525b18c2825141b118\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://206f6f19fe3197f5d6798cc86c5473bf2decb02a2860ee7ffb5efb5f60952e8c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bafaa9670b05671fb57b658c2960bf3ee32c84e52f86a3a3bbab01d5a38ff0c4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://773a0bc2aab73ffc2717e987543defc9b786cc22300aa54b4ba4f3454504e331\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-06T03:34:22Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1206 03:34:15.305330 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1206 03:34:15.307249 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1592846546/tls.crt::/tmp/serving-cert-1592846546/tls.key\\\\\\\"\\\\nI1206 03:34:21.814037 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1206 03:34:21.816756 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1206 03:34:21.816780 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1206 03:34:21.816806 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1206 03:34:21.816813 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1206 03:34:21.831301 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1206 03:34:21.831311 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1206 03:34:21.831324 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 03:34:21.831330 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 03:34:21.831335 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1206 03:34:21.831339 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1206 03:34:21.831342 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1206 03:34:21.831346 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1206 03:34:21.911269 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:04Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d40e0e7446f8416bf9da94fc53c01be937f9a51a11eb101157d2ddd4597720d2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:04Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2cba89f2f1bed96d59f183628d900d9b72c91eb472e273c8bed0babcab485910\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2cba89f2f1bed96d59f183628d900d9b72c91eb472e273c8bed0babcab485910\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:01Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:28Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:28 crc kubenswrapper[4980]: I1206 03:34:28.531430 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6ed68f66-1cf6-465a-8f66-76d5af6a44d8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://725e84f6bdfb722a8ce69158d03ea4c66a9f087642d55cb9651e003c81be0c8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://71f7b492ea07cdf7dc65fc48184ec1e64914eb1811bbfdddc0fb14c8ad6cb296\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b543aa4ee03d5a14909b67d876c5675c02e1df2a89a3c9ac8d3fb587fea53d8f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://32af3427e872a70cb3c2357116f04cabc8d3d9b69e7306b2e0de2ae22f9469d4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:01Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:28Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:28 crc kubenswrapper[4980]: I1206 03:34:28.545547 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:22Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:28Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:28 crc kubenswrapper[4980]: I1206 03:34:28.562675 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://334ea6a273fd7d268906c9f4a6dbf6f4df4fcdbf38f3d7d248cb7d6cf8537794\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ec44dcd5e4bf0a0466bb586ebd61259364dbbfe22499f4cbff854e3bd153082\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:28Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:28 crc kubenswrapper[4980]: I1206 03:34:28.573574 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:34:28 crc kubenswrapper[4980]: I1206 03:34:28.573639 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:34:28 crc kubenswrapper[4980]: I1206 03:34:28.573652 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:34:28 crc kubenswrapper[4980]: I1206 03:34:28.573673 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:34:28 crc kubenswrapper[4980]: I1206 03:34:28.573706 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:34:28Z","lastTransitionTime":"2025-12-06T03:34:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:34:28 crc kubenswrapper[4980]: I1206 03:34:28.578444 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:22Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:28Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:28 crc kubenswrapper[4980]: I1206 03:34:28.588488 4980 generic.go:334] "Generic (PLEG): container finished" podID="e5bac22a-512c-49d0-90db-47de6aaaefd0" containerID="c3b5dab476c7f227a6e4a2eae8469432667fb9fad345d9419ba59feab9fd385e" exitCode=0 Dec 06 03:34:28 crc kubenswrapper[4980]: I1206 03:34:28.588568 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-dx44q" event={"ID":"e5bac22a-512c-49d0-90db-47de6aaaefd0","Type":"ContainerDied","Data":"c3b5dab476c7f227a6e4a2eae8469432667fb9fad345d9419ba59feab9fd385e"} Dec 06 03:34:28 crc kubenswrapper[4980]: I1206 03:34:28.593220 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-sv5sz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"873977aa-4eb0-4526-bc0b-757ad2dea2a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01f0cc855c822a71ea8ef6a7ce6e25747ff9956691735973f935c4a28672a7a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lg7tx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:23Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-sv5sz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:28Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:28 crc kubenswrapper[4980]: I1206 03:34:28.616849 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5w4fr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5af0e768-3c9c-4401-ab56-57516cd3170e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:23Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:23Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfzrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfzrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfzrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfzrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfzrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfzrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfzrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfzrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d876df04b5ab0ba07a5cbcafbe4648403c5014c59eb7d40adee2657c3a255473\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d876df04b5ab0ba07a5cbcafbe4648403c5014c59eb7d40adee2657c3a255473\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfzrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:23Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-5w4fr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:28Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:28 crc kubenswrapper[4980]: I1206 03:34:28.634462 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cab608b3-a02f-475b-9aae-9b8f5d878de9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f7fcc44b42bcbb55509048e3a566bb557333c4a83e33d7b0fba8eb514317909\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c22e7dd8886c2dfbb742900ed86ec7b0c2edf9a1363e74f6666961d928f06c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://473e792b2d0a9644854ee0581c6120fc9c989bc342899ef72b1d056080fd4694\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ee8e936dfc39361d6f99460910591ca8c901a20f0d274bbb6aaebfa8a00ee06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://77bc12768b018625ba11a6bfd66688f926c4b7f3bc1d0e41c6551a18b485ecf3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://27cf07cd96893900da4afecd0d56fd9269bc60c1d23246b4685fcd8756357c19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://27cf07cd96893900da4afecd0d56fd9269bc60c1d23246b4685fcd8756357c19\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c49feb24566a8284378670fa0390710d38c237acc94ffb654f90f3fa4626e4c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c49feb24566a8284378670fa0390710d38c237acc94ffb654f90f3fa4626e4c8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:03Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://bb74d1c3d752a379fa394c450217badb45511fbcafad62f7b618131cab94b29e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bb74d1c3d752a379fa394c450217badb45511fbcafad62f7b618131cab94b29e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:01Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:28Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:28 crc kubenswrapper[4980]: I1206 03:34:28.645618 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://502c727d6baa31f400bee14c34b19abbe38af873f28f6a82a43547431755d981\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:28Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:28 crc kubenswrapper[4980]: I1206 03:34:28.658020 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-dx44q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5bac22a-512c-49d0-90db-47de6aaaefd0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:23Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:23Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:23Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f7mdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3b5dab476c7f227a6e4a2eae8469432667fb9fad345d9419ba59feab9fd385e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f7mdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f7mdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f7mdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f7mdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f7mdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f7mdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:23Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-dx44q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:28Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:28 crc kubenswrapper[4980]: I1206 03:34:28.672691 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-r5zfb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"320f44d1-a671-4a91-b328-a8b0fdd8f23a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://92393ecaaef8cfcfd01837671f8cc89803dd570b637b4a373b40b7f487788826\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zhm5t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52e8db644463794041978ced2c0591c13e0ada8e5f1a78a94a0a665218a37b5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zhm5t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:23Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-r5zfb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:28Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:28 crc kubenswrapper[4980]: I1206 03:34:28.676149 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:34:28 crc kubenswrapper[4980]: I1206 03:34:28.676187 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:34:28 crc kubenswrapper[4980]: I1206 03:34:28.676197 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:34:28 crc kubenswrapper[4980]: I1206 03:34:28.676229 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:34:28 crc kubenswrapper[4980]: I1206 03:34:28.676241 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:34:28Z","lastTransitionTime":"2025-12-06T03:34:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:34:28 crc kubenswrapper[4980]: I1206 03:34:28.684500 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:22Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:28Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:28 crc kubenswrapper[4980]: I1206 03:34:28.697920 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://334ea6a273fd7d268906c9f4a6dbf6f4df4fcdbf38f3d7d248cb7d6cf8537794\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ec44dcd5e4bf0a0466bb586ebd61259364dbbfe22499f4cbff854e3bd153082\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:28Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:28 crc kubenswrapper[4980]: I1206 03:34:28.707745 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:22Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:28Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:28 crc kubenswrapper[4980]: I1206 03:34:28.720613 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-sv5sz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"873977aa-4eb0-4526-bc0b-757ad2dea2a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01f0cc855c822a71ea8ef6a7ce6e25747ff9956691735973f935c4a28672a7a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lg7tx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:23Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-sv5sz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:28Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:28 crc kubenswrapper[4980]: I1206 03:34:28.742373 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5w4fr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5af0e768-3c9c-4401-ab56-57516cd3170e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:23Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:23Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfzrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfzrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfzrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfzrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfzrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfzrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfzrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfzrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d876df04b5ab0ba07a5cbcafbe4648403c5014c59eb7d40adee2657c3a255473\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d876df04b5ab0ba07a5cbcafbe4648403c5014c59eb7d40adee2657c3a255473\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfzrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:23Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-5w4fr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:28Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:28 crc kubenswrapper[4980]: I1206 03:34:28.759351 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cab608b3-a02f-475b-9aae-9b8f5d878de9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f7fcc44b42bcbb55509048e3a566bb557333c4a83e33d7b0fba8eb514317909\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c22e7dd8886c2dfbb742900ed86ec7b0c2edf9a1363e74f6666961d928f06c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://473e792b2d0a9644854ee0581c6120fc9c989bc342899ef72b1d056080fd4694\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ee8e936dfc39361d6f99460910591ca8c901a20f0d274bbb6aaebfa8a00ee06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://77bc12768b018625ba11a6bfd66688f926c4b7f3bc1d0e41c6551a18b485ecf3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://27cf07cd96893900da4afecd0d56fd9269bc60c1d23246b4685fcd8756357c19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://27cf07cd96893900da4afecd0d56fd9269bc60c1d23246b4685fcd8756357c19\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c49feb24566a8284378670fa0390710d38c237acc94ffb654f90f3fa4626e4c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c49feb24566a8284378670fa0390710d38c237acc94ffb654f90f3fa4626e4c8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:03Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://bb74d1c3d752a379fa394c450217badb45511fbcafad62f7b618131cab94b29e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bb74d1c3d752a379fa394c450217badb45511fbcafad62f7b618131cab94b29e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:01Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:28Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:28 crc kubenswrapper[4980]: I1206 03:34:28.768975 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://502c727d6baa31f400bee14c34b19abbe38af873f28f6a82a43547431755d981\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:28Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:28 crc kubenswrapper[4980]: I1206 03:34:28.779058 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:34:28 crc kubenswrapper[4980]: I1206 03:34:28.779114 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:34:28 crc kubenswrapper[4980]: I1206 03:34:28.779128 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:34:28 crc kubenswrapper[4980]: I1206 03:34:28.779147 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:34:28 crc kubenswrapper[4980]: I1206 03:34:28.779161 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:34:28Z","lastTransitionTime":"2025-12-06T03:34:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:34:28 crc kubenswrapper[4980]: I1206 03:34:28.783741 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-dx44q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5bac22a-512c-49d0-90db-47de6aaaefd0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:23Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:23Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:23Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f7mdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3b5dab476c7f227a6e4a2eae8469432667fb9fad345d9419ba59feab9fd385e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c3b5dab476c7f227a6e4a2eae8469432667fb9fad345d9419ba59feab9fd385e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f7mdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f7mdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f7mdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f7mdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f7mdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f7mdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:23Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-dx44q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:28Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:28 crc kubenswrapper[4980]: I1206 03:34:28.796657 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-r5zfb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"320f44d1-a671-4a91-b328-a8b0fdd8f23a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://92393ecaaef8cfcfd01837671f8cc89803dd570b637b4a373b40b7f487788826\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zhm5t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52e8db644463794041978ced2c0591c13e0ada8e5f1a78a94a0a665218a37b5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zhm5t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:23Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-r5zfb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:28Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:28 crc kubenswrapper[4980]: I1206 03:34:28.808540 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:28Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:28 crc kubenswrapper[4980]: I1206 03:34:28.819777 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4ef01e7d9809fd9cfc46e3a58b07159cdc6f5a3c210d1c9efc49bca27b30858\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:28Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:28 crc kubenswrapper[4980]: I1206 03:34:28.830913 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-kq4z8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f96c1f6c-55e8-4d86-bff3-5381581b7022\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://799498681afa7653408bc23b1e82d75c16e3c809d2ee7f322780e20c537eb710\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bdx7g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:23Z\\\"}}\" for pod \"openshift-multus\"/\"multus-kq4z8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:28Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:28 crc kubenswrapper[4980]: I1206 03:34:28.844852 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7f958150-549d-4c80-8a7e-23c39e4f9500\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c28b6895b14e73f4f3eb7a8ce93076c05092ba9b58ea0e3e879c65bcfa227a21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://76cff8a2d652df38b04eefc053832baf4cbb5db778ff5e525b18c2825141b118\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://206f6f19fe3197f5d6798cc86c5473bf2decb02a2860ee7ffb5efb5f60952e8c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bafaa9670b05671fb57b658c2960bf3ee32c84e52f86a3a3bbab01d5a38ff0c4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://773a0bc2aab73ffc2717e987543defc9b786cc22300aa54b4ba4f3454504e331\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-06T03:34:22Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1206 03:34:15.305330 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1206 03:34:15.307249 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1592846546/tls.crt::/tmp/serving-cert-1592846546/tls.key\\\\\\\"\\\\nI1206 03:34:21.814037 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1206 03:34:21.816756 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1206 03:34:21.816780 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1206 03:34:21.816806 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1206 03:34:21.816813 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1206 03:34:21.831301 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1206 03:34:21.831311 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1206 03:34:21.831324 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 03:34:21.831330 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 03:34:21.831335 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1206 03:34:21.831339 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1206 03:34:21.831342 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1206 03:34:21.831346 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1206 03:34:21.911269 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:04Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d40e0e7446f8416bf9da94fc53c01be937f9a51a11eb101157d2ddd4597720d2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:04Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2cba89f2f1bed96d59f183628d900d9b72c91eb472e273c8bed0babcab485910\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2cba89f2f1bed96d59f183628d900d9b72c91eb472e273c8bed0babcab485910\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:01Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:28Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:28 crc kubenswrapper[4980]: I1206 03:34:28.856378 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6ed68f66-1cf6-465a-8f66-76d5af6a44d8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://725e84f6bdfb722a8ce69158d03ea4c66a9f087642d55cb9651e003c81be0c8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://71f7b492ea07cdf7dc65fc48184ec1e64914eb1811bbfdddc0fb14c8ad6cb296\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b543aa4ee03d5a14909b67d876c5675c02e1df2a89a3c9ac8d3fb587fea53d8f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://32af3427e872a70cb3c2357116f04cabc8d3d9b69e7306b2e0de2ae22f9469d4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:01Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:28Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:28 crc kubenswrapper[4980]: I1206 03:34:28.881843 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:34:28 crc kubenswrapper[4980]: I1206 03:34:28.881875 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:34:28 crc kubenswrapper[4980]: I1206 03:34:28.881883 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:34:28 crc kubenswrapper[4980]: I1206 03:34:28.881898 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:34:28 crc kubenswrapper[4980]: I1206 03:34:28.881908 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:34:28Z","lastTransitionTime":"2025-12-06T03:34:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:34:28 crc kubenswrapper[4980]: I1206 03:34:28.984929 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:34:28 crc kubenswrapper[4980]: I1206 03:34:28.984980 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:34:28 crc kubenswrapper[4980]: I1206 03:34:28.984992 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:34:28 crc kubenswrapper[4980]: I1206 03:34:28.985011 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:34:28 crc kubenswrapper[4980]: I1206 03:34:28.985025 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:34:28Z","lastTransitionTime":"2025-12-06T03:34:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:34:29 crc kubenswrapper[4980]: I1206 03:34:29.088104 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:34:29 crc kubenswrapper[4980]: I1206 03:34:29.088153 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:34:29 crc kubenswrapper[4980]: I1206 03:34:29.088167 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:34:29 crc kubenswrapper[4980]: I1206 03:34:29.088184 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:34:29 crc kubenswrapper[4980]: I1206 03:34:29.088196 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:34:29Z","lastTransitionTime":"2025-12-06T03:34:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:34:29 crc kubenswrapper[4980]: I1206 03:34:29.183901 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 03:34:29 crc kubenswrapper[4980]: E1206 03:34:29.184072 4980 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 03:34:29 crc kubenswrapper[4980]: I1206 03:34:29.190645 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:34:29 crc kubenswrapper[4980]: I1206 03:34:29.190702 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:34:29 crc kubenswrapper[4980]: I1206 03:34:29.190725 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:34:29 crc kubenswrapper[4980]: I1206 03:34:29.190751 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:34:29 crc kubenswrapper[4980]: I1206 03:34:29.190769 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:34:29Z","lastTransitionTime":"2025-12-06T03:34:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:34:29 crc kubenswrapper[4980]: I1206 03:34:29.318252 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:34:29 crc kubenswrapper[4980]: I1206 03:34:29.318325 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:34:29 crc kubenswrapper[4980]: I1206 03:34:29.318343 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:34:29 crc kubenswrapper[4980]: I1206 03:34:29.318370 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:34:29 crc kubenswrapper[4980]: I1206 03:34:29.318788 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:34:29Z","lastTransitionTime":"2025-12-06T03:34:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:34:29 crc kubenswrapper[4980]: I1206 03:34:29.421730 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:34:29 crc kubenswrapper[4980]: I1206 03:34:29.421774 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:34:29 crc kubenswrapper[4980]: I1206 03:34:29.421783 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:34:29 crc kubenswrapper[4980]: I1206 03:34:29.421800 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:34:29 crc kubenswrapper[4980]: I1206 03:34:29.421812 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:34:29Z","lastTransitionTime":"2025-12-06T03:34:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:34:29 crc kubenswrapper[4980]: I1206 03:34:29.525156 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:34:29 crc kubenswrapper[4980]: I1206 03:34:29.525222 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:34:29 crc kubenswrapper[4980]: I1206 03:34:29.525241 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:34:29 crc kubenswrapper[4980]: I1206 03:34:29.525270 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:34:29 crc kubenswrapper[4980]: I1206 03:34:29.525287 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:34:29Z","lastTransitionTime":"2025-12-06T03:34:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:34:29 crc kubenswrapper[4980]: I1206 03:34:29.593886 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-dx44q" event={"ID":"e5bac22a-512c-49d0-90db-47de6aaaefd0","Type":"ContainerStarted","Data":"e921499231cbdf26b8918c98f9ad8c322fd7848865caeef048470de3e2c0efba"} Dec 06 03:34:29 crc kubenswrapper[4980]: I1206 03:34:29.612150 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6ed68f66-1cf6-465a-8f66-76d5af6a44d8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://725e84f6bdfb722a8ce69158d03ea4c66a9f087642d55cb9651e003c81be0c8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://71f7b492ea07cdf7dc65fc48184ec1e64914eb1811bbfdddc0fb14c8ad6cb296\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b543aa4ee03d5a14909b67d876c5675c02e1df2a89a3c9ac8d3fb587fea53d8f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://32af3427e872a70cb3c2357116f04cabc8d3d9b69e7306b2e0de2ae22f9469d4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:01Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:29Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:29 crc kubenswrapper[4980]: I1206 03:34:29.628434 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:34:29 crc kubenswrapper[4980]: I1206 03:34:29.628473 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:34:29 crc kubenswrapper[4980]: I1206 03:34:29.628485 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:34:29 crc kubenswrapper[4980]: I1206 03:34:29.628505 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:34:29 crc kubenswrapper[4980]: I1206 03:34:29.628547 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:34:29Z","lastTransitionTime":"2025-12-06T03:34:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:34:29 crc kubenswrapper[4980]: I1206 03:34:29.629473 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7f958150-549d-4c80-8a7e-23c39e4f9500\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c28b6895b14e73f4f3eb7a8ce93076c05092ba9b58ea0e3e879c65bcfa227a21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://76cff8a2d652df38b04eefc053832baf4cbb5db778ff5e525b18c2825141b118\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://206f6f19fe3197f5d6798cc86c5473bf2decb02a2860ee7ffb5efb5f60952e8c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bafaa9670b05671fb57b658c2960bf3ee32c84e52f86a3a3bbab01d5a38ff0c4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://773a0bc2aab73ffc2717e987543defc9b786cc22300aa54b4ba4f3454504e331\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-06T03:34:22Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1206 03:34:15.305330 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1206 03:34:15.307249 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1592846546/tls.crt::/tmp/serving-cert-1592846546/tls.key\\\\\\\"\\\\nI1206 03:34:21.814037 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1206 03:34:21.816756 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1206 03:34:21.816780 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1206 03:34:21.816806 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1206 03:34:21.816813 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1206 03:34:21.831301 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1206 03:34:21.831311 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1206 03:34:21.831324 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 03:34:21.831330 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 03:34:21.831335 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1206 03:34:21.831339 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1206 03:34:21.831342 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1206 03:34:21.831346 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1206 03:34:21.911269 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:04Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d40e0e7446f8416bf9da94fc53c01be937f9a51a11eb101157d2ddd4597720d2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:04Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2cba89f2f1bed96d59f183628d900d9b72c91eb472e273c8bed0babcab485910\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2cba89f2f1bed96d59f183628d900d9b72c91eb472e273c8bed0babcab485910\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:01Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:29Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:29 crc kubenswrapper[4980]: I1206 03:34:29.648371 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://334ea6a273fd7d268906c9f4a6dbf6f4df4fcdbf38f3d7d248cb7d6cf8537794\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ec44dcd5e4bf0a0466bb586ebd61259364dbbfe22499f4cbff854e3bd153082\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:29Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:29 crc kubenswrapper[4980]: I1206 03:34:29.667026 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:22Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:29Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:29 crc kubenswrapper[4980]: I1206 03:34:29.682294 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-sv5sz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"873977aa-4eb0-4526-bc0b-757ad2dea2a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01f0cc855c822a71ea8ef6a7ce6e25747ff9956691735973f935c4a28672a7a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lg7tx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:23Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-sv5sz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:29Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:29 crc kubenswrapper[4980]: I1206 03:34:29.704201 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5w4fr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5af0e768-3c9c-4401-ab56-57516cd3170e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:23Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:23Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfzrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfzrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfzrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfzrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfzrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfzrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfzrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfzrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d876df04b5ab0ba07a5cbcafbe4648403c5014c59eb7d40adee2657c3a255473\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d876df04b5ab0ba07a5cbcafbe4648403c5014c59eb7d40adee2657c3a255473\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfzrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:23Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-5w4fr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:29Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:29 crc kubenswrapper[4980]: I1206 03:34:29.717306 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:22Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:29Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:29 crc kubenswrapper[4980]: I1206 03:34:29.729035 4980 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/node-ca-cvzg8"] Dec 06 03:34:29 crc kubenswrapper[4980]: I1206 03:34:29.729394 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-cvzg8" Dec 06 03:34:29 crc kubenswrapper[4980]: I1206 03:34:29.731991 4980 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Dec 06 03:34:29 crc kubenswrapper[4980]: I1206 03:34:29.732153 4980 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Dec 06 03:34:29 crc kubenswrapper[4980]: I1206 03:34:29.732262 4980 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Dec 06 03:34:29 crc kubenswrapper[4980]: I1206 03:34:29.732577 4980 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Dec 06 03:34:29 crc kubenswrapper[4980]: I1206 03:34:29.733677 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:34:29 crc kubenswrapper[4980]: I1206 03:34:29.733735 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:34:29 crc kubenswrapper[4980]: I1206 03:34:29.733749 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:34:29 crc kubenswrapper[4980]: I1206 03:34:29.733779 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:34:29 crc kubenswrapper[4980]: I1206 03:34:29.733794 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:34:29Z","lastTransitionTime":"2025-12-06T03:34:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:34:29 crc kubenswrapper[4980]: I1206 03:34:29.735467 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-dx44q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5bac22a-512c-49d0-90db-47de6aaaefd0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:23Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:23Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:23Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f7mdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3b5dab476c7f227a6e4a2eae8469432667fb9fad345d9419ba59feab9fd385e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c3b5dab476c7f227a6e4a2eae8469432667fb9fad345d9419ba59feab9fd385e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f7mdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e921499231cbdf26b8918c98f9ad8c322fd7848865caeef048470de3e2c0efba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f7mdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f7mdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f7mdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f7mdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f7mdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:23Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-dx44q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:29Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:29 crc kubenswrapper[4980]: I1206 03:34:29.754021 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-r5zfb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"320f44d1-a671-4a91-b328-a8b0fdd8f23a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://92393ecaaef8cfcfd01837671f8cc89803dd570b637b4a373b40b7f487788826\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zhm5t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52e8db644463794041978ced2c0591c13e0ada8e5f1a78a94a0a665218a37b5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zhm5t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:23Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-r5zfb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:29Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:29 crc kubenswrapper[4980]: I1206 03:34:29.766181 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/f1cde706-268d-4d26-818f-2f42bfb37b5f-host\") pod \"node-ca-cvzg8\" (UID: \"f1cde706-268d-4d26-818f-2f42bfb37b5f\") " pod="openshift-image-registry/node-ca-cvzg8" Dec 06 03:34:29 crc kubenswrapper[4980]: I1206 03:34:29.766243 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rk86l\" (UniqueName: \"kubernetes.io/projected/f1cde706-268d-4d26-818f-2f42bfb37b5f-kube-api-access-rk86l\") pod \"node-ca-cvzg8\" (UID: \"f1cde706-268d-4d26-818f-2f42bfb37b5f\") " pod="openshift-image-registry/node-ca-cvzg8" Dec 06 03:34:29 crc kubenswrapper[4980]: I1206 03:34:29.766265 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/f1cde706-268d-4d26-818f-2f42bfb37b5f-serviceca\") pod \"node-ca-cvzg8\" (UID: \"f1cde706-268d-4d26-818f-2f42bfb37b5f\") " pod="openshift-image-registry/node-ca-cvzg8" Dec 06 03:34:29 crc kubenswrapper[4980]: I1206 03:34:29.792256 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cab608b3-a02f-475b-9aae-9b8f5d878de9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f7fcc44b42bcbb55509048e3a566bb557333c4a83e33d7b0fba8eb514317909\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c22e7dd8886c2dfbb742900ed86ec7b0c2edf9a1363e74f6666961d928f06c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://473e792b2d0a9644854ee0581c6120fc9c989bc342899ef72b1d056080fd4694\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ee8e936dfc39361d6f99460910591ca8c901a20f0d274bbb6aaebfa8a00ee06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://77bc12768b018625ba11a6bfd66688f926c4b7f3bc1d0e41c6551a18b485ecf3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://27cf07cd96893900da4afecd0d56fd9269bc60c1d23246b4685fcd8756357c19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://27cf07cd96893900da4afecd0d56fd9269bc60c1d23246b4685fcd8756357c19\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c49feb24566a8284378670fa0390710d38c237acc94ffb654f90f3fa4626e4c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c49feb24566a8284378670fa0390710d38c237acc94ffb654f90f3fa4626e4c8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:03Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://bb74d1c3d752a379fa394c450217badb45511fbcafad62f7b618131cab94b29e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bb74d1c3d752a379fa394c450217badb45511fbcafad62f7b618131cab94b29e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:01Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:29Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:29 crc kubenswrapper[4980]: I1206 03:34:29.807403 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://502c727d6baa31f400bee14c34b19abbe38af873f28f6a82a43547431755d981\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:29Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:29 crc kubenswrapper[4980]: I1206 03:34:29.821458 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-kq4z8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f96c1f6c-55e8-4d86-bff3-5381581b7022\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://799498681afa7653408bc23b1e82d75c16e3c809d2ee7f322780e20c537eb710\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bdx7g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:23Z\\\"}}\" for pod \"openshift-multus\"/\"multus-kq4z8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:29Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:29 crc kubenswrapper[4980]: I1206 03:34:29.835138 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:29Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:29 crc kubenswrapper[4980]: I1206 03:34:29.836999 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:34:29 crc kubenswrapper[4980]: I1206 03:34:29.837121 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:34:29 crc kubenswrapper[4980]: I1206 03:34:29.837183 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:34:29 crc kubenswrapper[4980]: I1206 03:34:29.837249 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:34:29 crc kubenswrapper[4980]: I1206 03:34:29.837311 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:34:29Z","lastTransitionTime":"2025-12-06T03:34:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:34:29 crc kubenswrapper[4980]: I1206 03:34:29.850547 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4ef01e7d9809fd9cfc46e3a58b07159cdc6f5a3c210d1c9efc49bca27b30858\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:29Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:29 crc kubenswrapper[4980]: I1206 03:34:29.867584 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/f1cde706-268d-4d26-818f-2f42bfb37b5f-host\") pod \"node-ca-cvzg8\" (UID: \"f1cde706-268d-4d26-818f-2f42bfb37b5f\") " pod="openshift-image-registry/node-ca-cvzg8" Dec 06 03:34:29 crc kubenswrapper[4980]: I1206 03:34:29.868008 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rk86l\" (UniqueName: \"kubernetes.io/projected/f1cde706-268d-4d26-818f-2f42bfb37b5f-kube-api-access-rk86l\") pod \"node-ca-cvzg8\" (UID: \"f1cde706-268d-4d26-818f-2f42bfb37b5f\") " pod="openshift-image-registry/node-ca-cvzg8" Dec 06 03:34:29 crc kubenswrapper[4980]: I1206 03:34:29.868102 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/f1cde706-268d-4d26-818f-2f42bfb37b5f-serviceca\") pod \"node-ca-cvzg8\" (UID: \"f1cde706-268d-4d26-818f-2f42bfb37b5f\") " pod="openshift-image-registry/node-ca-cvzg8" Dec 06 03:34:29 crc kubenswrapper[4980]: I1206 03:34:29.867710 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/f1cde706-268d-4d26-818f-2f42bfb37b5f-host\") pod \"node-ca-cvzg8\" (UID: \"f1cde706-268d-4d26-818f-2f42bfb37b5f\") " pod="openshift-image-registry/node-ca-cvzg8" Dec 06 03:34:29 crc kubenswrapper[4980]: I1206 03:34:29.869205 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/f1cde706-268d-4d26-818f-2f42bfb37b5f-serviceca\") pod \"node-ca-cvzg8\" (UID: \"f1cde706-268d-4d26-818f-2f42bfb37b5f\") " pod="openshift-image-registry/node-ca-cvzg8" Dec 06 03:34:29 crc kubenswrapper[4980]: I1206 03:34:29.877044 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:29Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:29 crc kubenswrapper[4980]: I1206 03:34:29.895140 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rk86l\" (UniqueName: \"kubernetes.io/projected/f1cde706-268d-4d26-818f-2f42bfb37b5f-kube-api-access-rk86l\") pod \"node-ca-cvzg8\" (UID: \"f1cde706-268d-4d26-818f-2f42bfb37b5f\") " pod="openshift-image-registry/node-ca-cvzg8" Dec 06 03:34:29 crc kubenswrapper[4980]: I1206 03:34:29.895571 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4ef01e7d9809fd9cfc46e3a58b07159cdc6f5a3c210d1c9efc49bca27b30858\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:29Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:29 crc kubenswrapper[4980]: I1206 03:34:29.914888 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-kq4z8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f96c1f6c-55e8-4d86-bff3-5381581b7022\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://799498681afa7653408bc23b1e82d75c16e3c809d2ee7f322780e20c537eb710\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bdx7g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:23Z\\\"}}\" for pod \"openshift-multus\"/\"multus-kq4z8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:29Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:29 crc kubenswrapper[4980]: I1206 03:34:29.932810 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7f958150-549d-4c80-8a7e-23c39e4f9500\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c28b6895b14e73f4f3eb7a8ce93076c05092ba9b58ea0e3e879c65bcfa227a21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://76cff8a2d652df38b04eefc053832baf4cbb5db778ff5e525b18c2825141b118\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://206f6f19fe3197f5d6798cc86c5473bf2decb02a2860ee7ffb5efb5f60952e8c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bafaa9670b05671fb57b658c2960bf3ee32c84e52f86a3a3bbab01d5a38ff0c4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://773a0bc2aab73ffc2717e987543defc9b786cc22300aa54b4ba4f3454504e331\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-06T03:34:22Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1206 03:34:15.305330 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1206 03:34:15.307249 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1592846546/tls.crt::/tmp/serving-cert-1592846546/tls.key\\\\\\\"\\\\nI1206 03:34:21.814037 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1206 03:34:21.816756 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1206 03:34:21.816780 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1206 03:34:21.816806 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1206 03:34:21.816813 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1206 03:34:21.831301 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1206 03:34:21.831311 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1206 03:34:21.831324 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 03:34:21.831330 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 03:34:21.831335 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1206 03:34:21.831339 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1206 03:34:21.831342 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1206 03:34:21.831346 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1206 03:34:21.911269 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:04Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d40e0e7446f8416bf9da94fc53c01be937f9a51a11eb101157d2ddd4597720d2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:04Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2cba89f2f1bed96d59f183628d900d9b72c91eb472e273c8bed0babcab485910\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2cba89f2f1bed96d59f183628d900d9b72c91eb472e273c8bed0babcab485910\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:01Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:29Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:29 crc kubenswrapper[4980]: I1206 03:34:29.940557 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:34:29 crc kubenswrapper[4980]: I1206 03:34:29.940620 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:34:29 crc kubenswrapper[4980]: I1206 03:34:29.940632 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:34:29 crc kubenswrapper[4980]: I1206 03:34:29.940667 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:34:29 crc kubenswrapper[4980]: I1206 03:34:29.940692 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:34:29Z","lastTransitionTime":"2025-12-06T03:34:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:34:29 crc kubenswrapper[4980]: I1206 03:34:29.952872 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6ed68f66-1cf6-465a-8f66-76d5af6a44d8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://725e84f6bdfb722a8ce69158d03ea4c66a9f087642d55cb9651e003c81be0c8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://71f7b492ea07cdf7dc65fc48184ec1e64914eb1811bbfdddc0fb14c8ad6cb296\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b543aa4ee03d5a14909b67d876c5675c02e1df2a89a3c9ac8d3fb587fea53d8f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://32af3427e872a70cb3c2357116f04cabc8d3d9b69e7306b2e0de2ae22f9469d4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:01Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:29Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:29 crc kubenswrapper[4980]: I1206 03:34:29.967791 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-sv5sz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"873977aa-4eb0-4526-bc0b-757ad2dea2a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01f0cc855c822a71ea8ef6a7ce6e25747ff9956691735973f935c4a28672a7a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lg7tx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:23Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-sv5sz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:29Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:29 crc kubenswrapper[4980]: I1206 03:34:29.968882 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 03:34:29 crc kubenswrapper[4980]: I1206 03:34:29.968961 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 03:34:29 crc kubenswrapper[4980]: E1206 03:34:29.969066 4980 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 06 03:34:29 crc kubenswrapper[4980]: E1206 03:34:29.969140 4980 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-06 03:34:37.969123927 +0000 UTC m=+37.213875198 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 06 03:34:29 crc kubenswrapper[4980]: E1206 03:34:29.969320 4980 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 06 03:34:29 crc kubenswrapper[4980]: E1206 03:34:29.969426 4980 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-06 03:34:37.969415755 +0000 UTC m=+37.214167026 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 06 03:34:29 crc kubenswrapper[4980]: I1206 03:34:29.986571 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5w4fr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5af0e768-3c9c-4401-ab56-57516cd3170e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:23Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:23Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfzrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfzrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfzrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfzrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfzrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfzrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfzrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfzrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d876df04b5ab0ba07a5cbcafbe4648403c5014c59eb7d40adee2657c3a255473\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d876df04b5ab0ba07a5cbcafbe4648403c5014c59eb7d40adee2657c3a255473\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfzrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:23Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-5w4fr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:29Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:30 crc kubenswrapper[4980]: I1206 03:34:30.001259 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:22Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:29Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:30 crc kubenswrapper[4980]: I1206 03:34:30.020861 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://334ea6a273fd7d268906c9f4a6dbf6f4df4fcdbf38f3d7d248cb7d6cf8537794\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ec44dcd5e4bf0a0466bb586ebd61259364dbbfe22499f4cbff854e3bd153082\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:30Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:30 crc kubenswrapper[4980]: I1206 03:34:30.033501 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:22Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:30Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:30 crc kubenswrapper[4980]: I1206 03:34:30.043649 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:34:30 crc kubenswrapper[4980]: I1206 03:34:30.043854 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:34:30 crc kubenswrapper[4980]: I1206 03:34:30.043919 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:34:30 crc kubenswrapper[4980]: I1206 03:34:30.043989 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:34:30 crc kubenswrapper[4980]: I1206 03:34:30.044052 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:34:30Z","lastTransitionTime":"2025-12-06T03:34:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:34:30 crc kubenswrapper[4980]: I1206 03:34:30.044792 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-cvzg8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f1cde706-268d-4d26-818f-2f42bfb37b5f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:29Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:29Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rk86l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:29Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-cvzg8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:30Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:30 crc kubenswrapper[4980]: I1206 03:34:30.047124 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-cvzg8" Dec 06 03:34:30 crc kubenswrapper[4980]: I1206 03:34:30.069691 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 03:34:30 crc kubenswrapper[4980]: I1206 03:34:30.069808 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 03:34:30 crc kubenswrapper[4980]: I1206 03:34:30.069864 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 03:34:30 crc kubenswrapper[4980]: E1206 03:34:30.070020 4980 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 06 03:34:30 crc kubenswrapper[4980]: E1206 03:34:30.070055 4980 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 06 03:34:30 crc kubenswrapper[4980]: E1206 03:34:30.070070 4980 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 06 03:34:30 crc kubenswrapper[4980]: E1206 03:34:30.070124 4980 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-06 03:34:38.070107549 +0000 UTC m=+37.314858820 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 06 03:34:30 crc kubenswrapper[4980]: E1206 03:34:30.070492 4980 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 03:34:38.070483319 +0000 UTC m=+37.315234590 (durationBeforeRetry 8s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 03:34:30 crc kubenswrapper[4980]: E1206 03:34:30.070573 4980 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 06 03:34:30 crc kubenswrapper[4980]: E1206 03:34:30.070588 4980 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 06 03:34:30 crc kubenswrapper[4980]: E1206 03:34:30.070600 4980 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 06 03:34:30 crc kubenswrapper[4980]: E1206 03:34:30.070622 4980 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-06 03:34:38.070615733 +0000 UTC m=+37.315367004 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 06 03:34:30 crc kubenswrapper[4980]: I1206 03:34:30.073948 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cab608b3-a02f-475b-9aae-9b8f5d878de9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f7fcc44b42bcbb55509048e3a566bb557333c4a83e33d7b0fba8eb514317909\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c22e7dd8886c2dfbb742900ed86ec7b0c2edf9a1363e74f6666961d928f06c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://473e792b2d0a9644854ee0581c6120fc9c989bc342899ef72b1d056080fd4694\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ee8e936dfc39361d6f99460910591ca8c901a20f0d274bbb6aaebfa8a00ee06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://77bc12768b018625ba11a6bfd66688f926c4b7f3bc1d0e41c6551a18b485ecf3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://27cf07cd96893900da4afecd0d56fd9269bc60c1d23246b4685fcd8756357c19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://27cf07cd96893900da4afecd0d56fd9269bc60c1d23246b4685fcd8756357c19\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c49feb24566a8284378670fa0390710d38c237acc94ffb654f90f3fa4626e4c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c49feb24566a8284378670fa0390710d38c237acc94ffb654f90f3fa4626e4c8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:03Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://bb74d1c3d752a379fa394c450217badb45511fbcafad62f7b618131cab94b29e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bb74d1c3d752a379fa394c450217badb45511fbcafad62f7b618131cab94b29e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:01Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:30Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:30 crc kubenswrapper[4980]: I1206 03:34:30.089937 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://502c727d6baa31f400bee14c34b19abbe38af873f28f6a82a43547431755d981\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:30Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:30 crc kubenswrapper[4980]: I1206 03:34:30.108618 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-dx44q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5bac22a-512c-49d0-90db-47de6aaaefd0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:23Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:23Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:23Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f7mdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3b5dab476c7f227a6e4a2eae8469432667fb9fad345d9419ba59feab9fd385e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c3b5dab476c7f227a6e4a2eae8469432667fb9fad345d9419ba59feab9fd385e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f7mdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e921499231cbdf26b8918c98f9ad8c322fd7848865caeef048470de3e2c0efba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f7mdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f7mdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f7mdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f7mdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f7mdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:23Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-dx44q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:30Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:30 crc kubenswrapper[4980]: I1206 03:34:30.125245 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-r5zfb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"320f44d1-a671-4a91-b328-a8b0fdd8f23a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://92393ecaaef8cfcfd01837671f8cc89803dd570b637b4a373b40b7f487788826\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zhm5t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52e8db644463794041978ced2c0591c13e0ada8e5f1a78a94a0a665218a37b5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zhm5t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:23Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-r5zfb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:30Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:30 crc kubenswrapper[4980]: I1206 03:34:30.147017 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:34:30 crc kubenswrapper[4980]: I1206 03:34:30.147065 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:34:30 crc kubenswrapper[4980]: I1206 03:34:30.147079 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:34:30 crc kubenswrapper[4980]: I1206 03:34:30.147100 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:34:30 crc kubenswrapper[4980]: I1206 03:34:30.147114 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:34:30Z","lastTransitionTime":"2025-12-06T03:34:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:34:30 crc kubenswrapper[4980]: I1206 03:34:30.184373 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 03:34:30 crc kubenswrapper[4980]: I1206 03:34:30.184454 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 03:34:30 crc kubenswrapper[4980]: E1206 03:34:30.184610 4980 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 03:34:30 crc kubenswrapper[4980]: E1206 03:34:30.185658 4980 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 03:34:30 crc kubenswrapper[4980]: I1206 03:34:30.250647 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:34:30 crc kubenswrapper[4980]: I1206 03:34:30.250691 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:34:30 crc kubenswrapper[4980]: I1206 03:34:30.250704 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:34:30 crc kubenswrapper[4980]: I1206 03:34:30.250723 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:34:30 crc kubenswrapper[4980]: I1206 03:34:30.250738 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:34:30Z","lastTransitionTime":"2025-12-06T03:34:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:34:30 crc kubenswrapper[4980]: I1206 03:34:30.353543 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:34:30 crc kubenswrapper[4980]: I1206 03:34:30.353584 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:34:30 crc kubenswrapper[4980]: I1206 03:34:30.353593 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:34:30 crc kubenswrapper[4980]: I1206 03:34:30.353611 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:34:30 crc kubenswrapper[4980]: I1206 03:34:30.353621 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:34:30Z","lastTransitionTime":"2025-12-06T03:34:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:34:30 crc kubenswrapper[4980]: I1206 03:34:30.457252 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:34:30 crc kubenswrapper[4980]: I1206 03:34:30.457329 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:34:30 crc kubenswrapper[4980]: I1206 03:34:30.457342 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:34:30 crc kubenswrapper[4980]: I1206 03:34:30.457363 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:34:30 crc kubenswrapper[4980]: I1206 03:34:30.457379 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:34:30Z","lastTransitionTime":"2025-12-06T03:34:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:34:30 crc kubenswrapper[4980]: I1206 03:34:30.560418 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:34:30 crc kubenswrapper[4980]: I1206 03:34:30.560457 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:34:30 crc kubenswrapper[4980]: I1206 03:34:30.560468 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:34:30 crc kubenswrapper[4980]: I1206 03:34:30.560484 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:34:30 crc kubenswrapper[4980]: I1206 03:34:30.560495 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:34:30Z","lastTransitionTime":"2025-12-06T03:34:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:34:30 crc kubenswrapper[4980]: I1206 03:34:30.603241 4980 generic.go:334] "Generic (PLEG): container finished" podID="e5bac22a-512c-49d0-90db-47de6aaaefd0" containerID="e921499231cbdf26b8918c98f9ad8c322fd7848865caeef048470de3e2c0efba" exitCode=0 Dec 06 03:34:30 crc kubenswrapper[4980]: I1206 03:34:30.603340 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-dx44q" event={"ID":"e5bac22a-512c-49d0-90db-47de6aaaefd0","Type":"ContainerDied","Data":"e921499231cbdf26b8918c98f9ad8c322fd7848865caeef048470de3e2c0efba"} Dec 06 03:34:30 crc kubenswrapper[4980]: I1206 03:34:30.615032 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-cvzg8" event={"ID":"f1cde706-268d-4d26-818f-2f42bfb37b5f","Type":"ContainerStarted","Data":"86f9a77acae360740be827eb0d1e1516e8a49cb9b197b6f5f9468ac70a101158"} Dec 06 03:34:30 crc kubenswrapper[4980]: I1206 03:34:30.615119 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-cvzg8" event={"ID":"f1cde706-268d-4d26-818f-2f42bfb37b5f","Type":"ContainerStarted","Data":"bf752c44c08a4099869261d58e83762c99e183e2938f9a11bab6eda431bb56ae"} Dec 06 03:34:30 crc kubenswrapper[4980]: I1206 03:34:30.635119 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7f958150-549d-4c80-8a7e-23c39e4f9500\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c28b6895b14e73f4f3eb7a8ce93076c05092ba9b58ea0e3e879c65bcfa227a21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://76cff8a2d652df38b04eefc053832baf4cbb5db778ff5e525b18c2825141b118\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://206f6f19fe3197f5d6798cc86c5473bf2decb02a2860ee7ffb5efb5f60952e8c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bafaa9670b05671fb57b658c2960bf3ee32c84e52f86a3a3bbab01d5a38ff0c4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://773a0bc2aab73ffc2717e987543defc9b786cc22300aa54b4ba4f3454504e331\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-06T03:34:22Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1206 03:34:15.305330 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1206 03:34:15.307249 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1592846546/tls.crt::/tmp/serving-cert-1592846546/tls.key\\\\\\\"\\\\nI1206 03:34:21.814037 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1206 03:34:21.816756 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1206 03:34:21.816780 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1206 03:34:21.816806 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1206 03:34:21.816813 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1206 03:34:21.831301 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1206 03:34:21.831311 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1206 03:34:21.831324 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 03:34:21.831330 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 03:34:21.831335 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1206 03:34:21.831339 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1206 03:34:21.831342 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1206 03:34:21.831346 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1206 03:34:21.911269 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:04Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d40e0e7446f8416bf9da94fc53c01be937f9a51a11eb101157d2ddd4597720d2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:04Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2cba89f2f1bed96d59f183628d900d9b72c91eb472e273c8bed0babcab485910\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2cba89f2f1bed96d59f183628d900d9b72c91eb472e273c8bed0babcab485910\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:01Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:30Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:30 crc kubenswrapper[4980]: I1206 03:34:30.647340 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5w4fr" event={"ID":"5af0e768-3c9c-4401-ab56-57516cd3170e","Type":"ContainerStarted","Data":"fe261f4322b85702181d7473ab2a9a4d2d4b869d586c3091f811b12a81a63aae"} Dec 06 03:34:30 crc kubenswrapper[4980]: I1206 03:34:30.666762 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:34:30 crc kubenswrapper[4980]: I1206 03:34:30.666846 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:34:30 crc kubenswrapper[4980]: I1206 03:34:30.666863 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:34:30 crc kubenswrapper[4980]: I1206 03:34:30.666886 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:34:30 crc kubenswrapper[4980]: I1206 03:34:30.666898 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:34:30Z","lastTransitionTime":"2025-12-06T03:34:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:34:30 crc kubenswrapper[4980]: I1206 03:34:30.669390 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6ed68f66-1cf6-465a-8f66-76d5af6a44d8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://725e84f6bdfb722a8ce69158d03ea4c66a9f087642d55cb9651e003c81be0c8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://71f7b492ea07cdf7dc65fc48184ec1e64914eb1811bbfdddc0fb14c8ad6cb296\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b543aa4ee03d5a14909b67d876c5675c02e1df2a89a3c9ac8d3fb587fea53d8f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://32af3427e872a70cb3c2357116f04cabc8d3d9b69e7306b2e0de2ae22f9469d4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:01Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:30Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:30 crc kubenswrapper[4980]: I1206 03:34:30.710893 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:22Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:30Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:30 crc kubenswrapper[4980]: I1206 03:34:30.736439 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://334ea6a273fd7d268906c9f4a6dbf6f4df4fcdbf38f3d7d248cb7d6cf8537794\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ec44dcd5e4bf0a0466bb586ebd61259364dbbfe22499f4cbff854e3bd153082\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:30Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:30 crc kubenswrapper[4980]: I1206 03:34:30.751736 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:22Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:30Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:30 crc kubenswrapper[4980]: I1206 03:34:30.763706 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-sv5sz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"873977aa-4eb0-4526-bc0b-757ad2dea2a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01f0cc855c822a71ea8ef6a7ce6e25747ff9956691735973f935c4a28672a7a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lg7tx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:23Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-sv5sz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:30Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:30 crc kubenswrapper[4980]: I1206 03:34:30.771230 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:34:30 crc kubenswrapper[4980]: I1206 03:34:30.771269 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:34:30 crc kubenswrapper[4980]: I1206 03:34:30.771277 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:34:30 crc kubenswrapper[4980]: I1206 03:34:30.771308 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:34:30 crc kubenswrapper[4980]: I1206 03:34:30.771318 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:34:30Z","lastTransitionTime":"2025-12-06T03:34:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:34:30 crc kubenswrapper[4980]: I1206 03:34:30.787917 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5w4fr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5af0e768-3c9c-4401-ab56-57516cd3170e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:23Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:23Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfzrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfzrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfzrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfzrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfzrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfzrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfzrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfzrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d876df04b5ab0ba07a5cbcafbe4648403c5014c59eb7d40adee2657c3a255473\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d876df04b5ab0ba07a5cbcafbe4648403c5014c59eb7d40adee2657c3a255473\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfzrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:23Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-5w4fr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:30Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:30 crc kubenswrapper[4980]: I1206 03:34:30.809018 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cab608b3-a02f-475b-9aae-9b8f5d878de9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f7fcc44b42bcbb55509048e3a566bb557333c4a83e33d7b0fba8eb514317909\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c22e7dd8886c2dfbb742900ed86ec7b0c2edf9a1363e74f6666961d928f06c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://473e792b2d0a9644854ee0581c6120fc9c989bc342899ef72b1d056080fd4694\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ee8e936dfc39361d6f99460910591ca8c901a20f0d274bbb6aaebfa8a00ee06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://77bc12768b018625ba11a6bfd66688f926c4b7f3bc1d0e41c6551a18b485ecf3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://27cf07cd96893900da4afecd0d56fd9269bc60c1d23246b4685fcd8756357c19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://27cf07cd96893900da4afecd0d56fd9269bc60c1d23246b4685fcd8756357c19\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c49feb24566a8284378670fa0390710d38c237acc94ffb654f90f3fa4626e4c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c49feb24566a8284378670fa0390710d38c237acc94ffb654f90f3fa4626e4c8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:03Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://bb74d1c3d752a379fa394c450217badb45511fbcafad62f7b618131cab94b29e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bb74d1c3d752a379fa394c450217badb45511fbcafad62f7b618131cab94b29e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:01Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:30Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:30 crc kubenswrapper[4980]: I1206 03:34:30.849764 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://502c727d6baa31f400bee14c34b19abbe38af873f28f6a82a43547431755d981\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:30Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:30 crc kubenswrapper[4980]: I1206 03:34:30.863320 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-dx44q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5bac22a-512c-49d0-90db-47de6aaaefd0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:23Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:23Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:23Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f7mdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3b5dab476c7f227a6e4a2eae8469432667fb9fad345d9419ba59feab9fd385e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c3b5dab476c7f227a6e4a2eae8469432667fb9fad345d9419ba59feab9fd385e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f7mdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e921499231cbdf26b8918c98f9ad8c322fd7848865caeef048470de3e2c0efba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e921499231cbdf26b8918c98f9ad8c322fd7848865caeef048470de3e2c0efba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f7mdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f7mdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f7mdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f7mdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f7mdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:23Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-dx44q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:30Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:30 crc kubenswrapper[4980]: I1206 03:34:30.874470 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:34:30 crc kubenswrapper[4980]: I1206 03:34:30.874530 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:34:30 crc kubenswrapper[4980]: I1206 03:34:30.874542 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:34:30 crc kubenswrapper[4980]: I1206 03:34:30.874562 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:34:30 crc kubenswrapper[4980]: I1206 03:34:30.874574 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:34:30Z","lastTransitionTime":"2025-12-06T03:34:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:34:30 crc kubenswrapper[4980]: I1206 03:34:30.876262 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-r5zfb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"320f44d1-a671-4a91-b328-a8b0fdd8f23a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://92393ecaaef8cfcfd01837671f8cc89803dd570b637b4a373b40b7f487788826\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zhm5t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52e8db644463794041978ced2c0591c13e0ada8e5f1a78a94a0a665218a37b5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zhm5t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:23Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-r5zfb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:30Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:30 crc kubenswrapper[4980]: I1206 03:34:30.886448 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-cvzg8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f1cde706-268d-4d26-818f-2f42bfb37b5f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:29Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:29Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rk86l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:29Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-cvzg8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:30Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:30 crc kubenswrapper[4980]: I1206 03:34:30.899818 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:30Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:30 crc kubenswrapper[4980]: I1206 03:34:30.913253 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4ef01e7d9809fd9cfc46e3a58b07159cdc6f5a3c210d1c9efc49bca27b30858\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:30Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:30 crc kubenswrapper[4980]: I1206 03:34:30.977254 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:34:30 crc kubenswrapper[4980]: I1206 03:34:30.977288 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:34:30 crc kubenswrapper[4980]: I1206 03:34:30.977299 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:34:30 crc kubenswrapper[4980]: I1206 03:34:30.977316 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:34:30 crc kubenswrapper[4980]: I1206 03:34:30.977325 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:34:30Z","lastTransitionTime":"2025-12-06T03:34:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:34:31 crc kubenswrapper[4980]: I1206 03:34:31.005197 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-kq4z8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f96c1f6c-55e8-4d86-bff3-5381581b7022\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://799498681afa7653408bc23b1e82d75c16e3c809d2ee7f322780e20c537eb710\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bdx7g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:23Z\\\"}}\" for pod \"openshift-multus\"/\"multus-kq4z8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:31Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:31 crc kubenswrapper[4980]: I1206 03:34:31.017624 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6ed68f66-1cf6-465a-8f66-76d5af6a44d8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://725e84f6bdfb722a8ce69158d03ea4c66a9f087642d55cb9651e003c81be0c8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://71f7b492ea07cdf7dc65fc48184ec1e64914eb1811bbfdddc0fb14c8ad6cb296\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b543aa4ee03d5a14909b67d876c5675c02e1df2a89a3c9ac8d3fb587fea53d8f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://32af3427e872a70cb3c2357116f04cabc8d3d9b69e7306b2e0de2ae22f9469d4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:01Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:31Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:31 crc kubenswrapper[4980]: I1206 03:34:31.030905 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7f958150-549d-4c80-8a7e-23c39e4f9500\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c28b6895b14e73f4f3eb7a8ce93076c05092ba9b58ea0e3e879c65bcfa227a21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://76cff8a2d652df38b04eefc053832baf4cbb5db778ff5e525b18c2825141b118\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://206f6f19fe3197f5d6798cc86c5473bf2decb02a2860ee7ffb5efb5f60952e8c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bafaa9670b05671fb57b658c2960bf3ee32c84e52f86a3a3bbab01d5a38ff0c4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://773a0bc2aab73ffc2717e987543defc9b786cc22300aa54b4ba4f3454504e331\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-06T03:34:22Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1206 03:34:15.305330 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1206 03:34:15.307249 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1592846546/tls.crt::/tmp/serving-cert-1592846546/tls.key\\\\\\\"\\\\nI1206 03:34:21.814037 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1206 03:34:21.816756 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1206 03:34:21.816780 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1206 03:34:21.816806 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1206 03:34:21.816813 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1206 03:34:21.831301 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1206 03:34:21.831311 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1206 03:34:21.831324 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 03:34:21.831330 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 03:34:21.831335 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1206 03:34:21.831339 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1206 03:34:21.831342 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1206 03:34:21.831346 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1206 03:34:21.911269 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:04Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d40e0e7446f8416bf9da94fc53c01be937f9a51a11eb101157d2ddd4597720d2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:04Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2cba89f2f1bed96d59f183628d900d9b72c91eb472e273c8bed0babcab485910\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2cba89f2f1bed96d59f183628d900d9b72c91eb472e273c8bed0babcab485910\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:01Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:31Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:31 crc kubenswrapper[4980]: I1206 03:34:31.044888 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://334ea6a273fd7d268906c9f4a6dbf6f4df4fcdbf38f3d7d248cb7d6cf8537794\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ec44dcd5e4bf0a0466bb586ebd61259364dbbfe22499f4cbff854e3bd153082\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:31Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:31 crc kubenswrapper[4980]: I1206 03:34:31.057927 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:22Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:31Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:31 crc kubenswrapper[4980]: I1206 03:34:31.070900 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-sv5sz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"873977aa-4eb0-4526-bc0b-757ad2dea2a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01f0cc855c822a71ea8ef6a7ce6e25747ff9956691735973f935c4a28672a7a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lg7tx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:23Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-sv5sz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:31Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:31 crc kubenswrapper[4980]: I1206 03:34:31.087719 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:34:31 crc kubenswrapper[4980]: I1206 03:34:31.087766 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:34:31 crc kubenswrapper[4980]: I1206 03:34:31.087777 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:34:31 crc kubenswrapper[4980]: I1206 03:34:31.087799 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:34:31 crc kubenswrapper[4980]: I1206 03:34:31.087817 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:34:31Z","lastTransitionTime":"2025-12-06T03:34:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:34:31 crc kubenswrapper[4980]: I1206 03:34:31.103106 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5w4fr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5af0e768-3c9c-4401-ab56-57516cd3170e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:23Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:23Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfzrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfzrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfzrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfzrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfzrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfzrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfzrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfzrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d876df04b5ab0ba07a5cbcafbe4648403c5014c59eb7d40adee2657c3a255473\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d876df04b5ab0ba07a5cbcafbe4648403c5014c59eb7d40adee2657c3a255473\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfzrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:23Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-5w4fr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:31Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:31 crc kubenswrapper[4980]: I1206 03:34:31.122031 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:22Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:31Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:31 crc kubenswrapper[4980]: I1206 03:34:31.141306 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-dx44q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5bac22a-512c-49d0-90db-47de6aaaefd0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:23Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:23Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:23Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f7mdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3b5dab476c7f227a6e4a2eae8469432667fb9fad345d9419ba59feab9fd385e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c3b5dab476c7f227a6e4a2eae8469432667fb9fad345d9419ba59feab9fd385e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f7mdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e921499231cbdf26b8918c98f9ad8c322fd7848865caeef048470de3e2c0efba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e921499231cbdf26b8918c98f9ad8c322fd7848865caeef048470de3e2c0efba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f7mdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f7mdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f7mdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f7mdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f7mdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:23Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-dx44q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:31Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:31 crc kubenswrapper[4980]: I1206 03:34:31.155870 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-r5zfb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"320f44d1-a671-4a91-b328-a8b0fdd8f23a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://92393ecaaef8cfcfd01837671f8cc89803dd570b637b4a373b40b7f487788826\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zhm5t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52e8db644463794041978ced2c0591c13e0ada8e5f1a78a94a0a665218a37b5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zhm5t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:23Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-r5zfb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:31Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:31 crc kubenswrapper[4980]: I1206 03:34:31.166856 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-cvzg8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f1cde706-268d-4d26-818f-2f42bfb37b5f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://86f9a77acae360740be827eb0d1e1516e8a49cb9b197b6f5f9468ac70a101158\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rk86l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:29Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-cvzg8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:31Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:31 crc kubenswrapper[4980]: I1206 03:34:31.184279 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 03:34:31 crc kubenswrapper[4980]: E1206 03:34:31.184407 4980 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 03:34:31 crc kubenswrapper[4980]: I1206 03:34:31.189744 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:34:31 crc kubenswrapper[4980]: I1206 03:34:31.189775 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:34:31 crc kubenswrapper[4980]: I1206 03:34:31.189789 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:34:31 crc kubenswrapper[4980]: I1206 03:34:31.189809 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:34:31 crc kubenswrapper[4980]: I1206 03:34:31.189823 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:34:31Z","lastTransitionTime":"2025-12-06T03:34:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:34:31 crc kubenswrapper[4980]: I1206 03:34:31.199634 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cab608b3-a02f-475b-9aae-9b8f5d878de9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f7fcc44b42bcbb55509048e3a566bb557333c4a83e33d7b0fba8eb514317909\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c22e7dd8886c2dfbb742900ed86ec7b0c2edf9a1363e74f6666961d928f06c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://473e792b2d0a9644854ee0581c6120fc9c989bc342899ef72b1d056080fd4694\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ee8e936dfc39361d6f99460910591ca8c901a20f0d274bbb6aaebfa8a00ee06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://77bc12768b018625ba11a6bfd66688f926c4b7f3bc1d0e41c6551a18b485ecf3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://27cf07cd96893900da4afecd0d56fd9269bc60c1d23246b4685fcd8756357c19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://27cf07cd96893900da4afecd0d56fd9269bc60c1d23246b4685fcd8756357c19\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c49feb24566a8284378670fa0390710d38c237acc94ffb654f90f3fa4626e4c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c49feb24566a8284378670fa0390710d38c237acc94ffb654f90f3fa4626e4c8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:03Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://bb74d1c3d752a379fa394c450217badb45511fbcafad62f7b618131cab94b29e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bb74d1c3d752a379fa394c450217badb45511fbcafad62f7b618131cab94b29e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:01Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:31Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:31 crc kubenswrapper[4980]: I1206 03:34:31.217633 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://502c727d6baa31f400bee14c34b19abbe38af873f28f6a82a43547431755d981\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:31Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:31 crc kubenswrapper[4980]: I1206 03:34:31.230862 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-kq4z8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f96c1f6c-55e8-4d86-bff3-5381581b7022\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://799498681afa7653408bc23b1e82d75c16e3c809d2ee7f322780e20c537eb710\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bdx7g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:23Z\\\"}}\" for pod \"openshift-multus\"/\"multus-kq4z8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:31Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:31 crc kubenswrapper[4980]: I1206 03:34:31.246470 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:31Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:31 crc kubenswrapper[4980]: I1206 03:34:31.260245 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4ef01e7d9809fd9cfc46e3a58b07159cdc6f5a3c210d1c9efc49bca27b30858\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:31Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:31 crc kubenswrapper[4980]: I1206 03:34:31.274424 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:31Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:31 crc kubenswrapper[4980]: I1206 03:34:31.344066 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:34:31 crc kubenswrapper[4980]: I1206 03:34:31.344137 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:34:31 crc kubenswrapper[4980]: I1206 03:34:31.344156 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:34:31 crc kubenswrapper[4980]: I1206 03:34:31.344184 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:34:31 crc kubenswrapper[4980]: I1206 03:34:31.344200 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:34:31Z","lastTransitionTime":"2025-12-06T03:34:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:34:31 crc kubenswrapper[4980]: I1206 03:34:31.360677 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4ef01e7d9809fd9cfc46e3a58b07159cdc6f5a3c210d1c9efc49bca27b30858\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:31Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:31 crc kubenswrapper[4980]: I1206 03:34:31.377975 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-kq4z8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f96c1f6c-55e8-4d86-bff3-5381581b7022\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://799498681afa7653408bc23b1e82d75c16e3c809d2ee7f322780e20c537eb710\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bdx7g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:23Z\\\"}}\" for pod \"openshift-multus\"/\"multus-kq4z8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:31Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:31 crc kubenswrapper[4980]: I1206 03:34:31.393355 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7f958150-549d-4c80-8a7e-23c39e4f9500\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c28b6895b14e73f4f3eb7a8ce93076c05092ba9b58ea0e3e879c65bcfa227a21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://76cff8a2d652df38b04eefc053832baf4cbb5db778ff5e525b18c2825141b118\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://206f6f19fe3197f5d6798cc86c5473bf2decb02a2860ee7ffb5efb5f60952e8c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bafaa9670b05671fb57b658c2960bf3ee32c84e52f86a3a3bbab01d5a38ff0c4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://773a0bc2aab73ffc2717e987543defc9b786cc22300aa54b4ba4f3454504e331\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-06T03:34:22Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1206 03:34:15.305330 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1206 03:34:15.307249 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1592846546/tls.crt::/tmp/serving-cert-1592846546/tls.key\\\\\\\"\\\\nI1206 03:34:21.814037 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1206 03:34:21.816756 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1206 03:34:21.816780 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1206 03:34:21.816806 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1206 03:34:21.816813 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1206 03:34:21.831301 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1206 03:34:21.831311 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1206 03:34:21.831324 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 03:34:21.831330 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 03:34:21.831335 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1206 03:34:21.831339 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1206 03:34:21.831342 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1206 03:34:21.831346 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1206 03:34:21.911269 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:04Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d40e0e7446f8416bf9da94fc53c01be937f9a51a11eb101157d2ddd4597720d2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:04Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2cba89f2f1bed96d59f183628d900d9b72c91eb472e273c8bed0babcab485910\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2cba89f2f1bed96d59f183628d900d9b72c91eb472e273c8bed0babcab485910\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:01Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:31Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:31 crc kubenswrapper[4980]: I1206 03:34:31.407591 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6ed68f66-1cf6-465a-8f66-76d5af6a44d8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://725e84f6bdfb722a8ce69158d03ea4c66a9f087642d55cb9651e003c81be0c8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://71f7b492ea07cdf7dc65fc48184ec1e64914eb1811bbfdddc0fb14c8ad6cb296\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b543aa4ee03d5a14909b67d876c5675c02e1df2a89a3c9ac8d3fb587fea53d8f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://32af3427e872a70cb3c2357116f04cabc8d3d9b69e7306b2e0de2ae22f9469d4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:01Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:31Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:31 crc kubenswrapper[4980]: I1206 03:34:31.421557 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:22Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:31Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:31 crc kubenswrapper[4980]: I1206 03:34:31.435725 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://334ea6a273fd7d268906c9f4a6dbf6f4df4fcdbf38f3d7d248cb7d6cf8537794\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ec44dcd5e4bf0a0466bb586ebd61259364dbbfe22499f4cbff854e3bd153082\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:31Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:31 crc kubenswrapper[4980]: I1206 03:34:31.446634 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:34:31 crc kubenswrapper[4980]: I1206 03:34:31.446684 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:34:31 crc kubenswrapper[4980]: I1206 03:34:31.446696 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:34:31 crc kubenswrapper[4980]: I1206 03:34:31.446712 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:34:31 crc kubenswrapper[4980]: I1206 03:34:31.446721 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:34:31Z","lastTransitionTime":"2025-12-06T03:34:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:34:31 crc kubenswrapper[4980]: I1206 03:34:31.449402 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:22Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:31Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:31 crc kubenswrapper[4980]: I1206 03:34:31.459696 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-sv5sz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"873977aa-4eb0-4526-bc0b-757ad2dea2a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01f0cc855c822a71ea8ef6a7ce6e25747ff9956691735973f935c4a28672a7a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lg7tx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:23Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-sv5sz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:31Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:31 crc kubenswrapper[4980]: I1206 03:34:31.486133 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5w4fr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5af0e768-3c9c-4401-ab56-57516cd3170e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:23Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:23Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfzrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfzrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfzrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfzrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfzrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfzrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfzrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfzrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d876df04b5ab0ba07a5cbcafbe4648403c5014c59eb7d40adee2657c3a255473\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d876df04b5ab0ba07a5cbcafbe4648403c5014c59eb7d40adee2657c3a255473\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfzrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:23Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-5w4fr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:31Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:31 crc kubenswrapper[4980]: I1206 03:34:31.506777 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cab608b3-a02f-475b-9aae-9b8f5d878de9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f7fcc44b42bcbb55509048e3a566bb557333c4a83e33d7b0fba8eb514317909\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c22e7dd8886c2dfbb742900ed86ec7b0c2edf9a1363e74f6666961d928f06c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://473e792b2d0a9644854ee0581c6120fc9c989bc342899ef72b1d056080fd4694\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ee8e936dfc39361d6f99460910591ca8c901a20f0d274bbb6aaebfa8a00ee06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://77bc12768b018625ba11a6bfd66688f926c4b7f3bc1d0e41c6551a18b485ecf3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://27cf07cd96893900da4afecd0d56fd9269bc60c1d23246b4685fcd8756357c19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://27cf07cd96893900da4afecd0d56fd9269bc60c1d23246b4685fcd8756357c19\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c49feb24566a8284378670fa0390710d38c237acc94ffb654f90f3fa4626e4c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c49feb24566a8284378670fa0390710d38c237acc94ffb654f90f3fa4626e4c8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:03Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://bb74d1c3d752a379fa394c450217badb45511fbcafad62f7b618131cab94b29e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bb74d1c3d752a379fa394c450217badb45511fbcafad62f7b618131cab94b29e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:01Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:31Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:31 crc kubenswrapper[4980]: I1206 03:34:31.520109 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://502c727d6baa31f400bee14c34b19abbe38af873f28f6a82a43547431755d981\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:31Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:31 crc kubenswrapper[4980]: I1206 03:34:31.534547 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-dx44q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5bac22a-512c-49d0-90db-47de6aaaefd0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:23Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:23Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:23Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f7mdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3b5dab476c7f227a6e4a2eae8469432667fb9fad345d9419ba59feab9fd385e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c3b5dab476c7f227a6e4a2eae8469432667fb9fad345d9419ba59feab9fd385e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f7mdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e921499231cbdf26b8918c98f9ad8c322fd7848865caeef048470de3e2c0efba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e921499231cbdf26b8918c98f9ad8c322fd7848865caeef048470de3e2c0efba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f7mdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f7mdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f7mdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f7mdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f7mdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:23Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-dx44q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:31Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:31 crc kubenswrapper[4980]: I1206 03:34:31.547751 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-r5zfb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"320f44d1-a671-4a91-b328-a8b0fdd8f23a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://92393ecaaef8cfcfd01837671f8cc89803dd570b637b4a373b40b7f487788826\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zhm5t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52e8db644463794041978ced2c0591c13e0ada8e5f1a78a94a0a665218a37b5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zhm5t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:23Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-r5zfb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:31Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:31 crc kubenswrapper[4980]: I1206 03:34:31.549498 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:34:31 crc kubenswrapper[4980]: I1206 03:34:31.549545 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:34:31 crc kubenswrapper[4980]: I1206 03:34:31.549555 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:34:31 crc kubenswrapper[4980]: I1206 03:34:31.549572 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:34:31 crc kubenswrapper[4980]: I1206 03:34:31.549584 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:34:31Z","lastTransitionTime":"2025-12-06T03:34:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:34:31 crc kubenswrapper[4980]: I1206 03:34:31.559952 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-cvzg8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f1cde706-268d-4d26-818f-2f42bfb37b5f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://86f9a77acae360740be827eb0d1e1516e8a49cb9b197b6f5f9468ac70a101158\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rk86l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:29Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-cvzg8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:31Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:31 crc kubenswrapper[4980]: I1206 03:34:31.652574 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:34:31 crc kubenswrapper[4980]: I1206 03:34:31.652604 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:34:31 crc kubenswrapper[4980]: I1206 03:34:31.652614 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:34:31 crc kubenswrapper[4980]: I1206 03:34:31.652627 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:34:31 crc kubenswrapper[4980]: I1206 03:34:31.652635 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:34:31Z","lastTransitionTime":"2025-12-06T03:34:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:34:31 crc kubenswrapper[4980]: I1206 03:34:31.653580 4980 generic.go:334] "Generic (PLEG): container finished" podID="e5bac22a-512c-49d0-90db-47de6aaaefd0" containerID="7cbfe7d8cdb89916c238296592fab45b9e92e2c585baa6d67230decd1c84c5bb" exitCode=0 Dec 06 03:34:31 crc kubenswrapper[4980]: I1206 03:34:31.653631 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-dx44q" event={"ID":"e5bac22a-512c-49d0-90db-47de6aaaefd0","Type":"ContainerDied","Data":"7cbfe7d8cdb89916c238296592fab45b9e92e2c585baa6d67230decd1c84c5bb"} Dec 06 03:34:31 crc kubenswrapper[4980]: I1206 03:34:31.668823 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:31Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:31 crc kubenswrapper[4980]: I1206 03:34:31.684866 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4ef01e7d9809fd9cfc46e3a58b07159cdc6f5a3c210d1c9efc49bca27b30858\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:31Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:31 crc kubenswrapper[4980]: I1206 03:34:31.696073 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-kq4z8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f96c1f6c-55e8-4d86-bff3-5381581b7022\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://799498681afa7653408bc23b1e82d75c16e3c809d2ee7f322780e20c537eb710\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bdx7g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:23Z\\\"}}\" for pod \"openshift-multus\"/\"multus-kq4z8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:31Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:31 crc kubenswrapper[4980]: I1206 03:34:31.712198 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7f958150-549d-4c80-8a7e-23c39e4f9500\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c28b6895b14e73f4f3eb7a8ce93076c05092ba9b58ea0e3e879c65bcfa227a21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://76cff8a2d652df38b04eefc053832baf4cbb5db778ff5e525b18c2825141b118\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://206f6f19fe3197f5d6798cc86c5473bf2decb02a2860ee7ffb5efb5f60952e8c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bafaa9670b05671fb57b658c2960bf3ee32c84e52f86a3a3bbab01d5a38ff0c4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://773a0bc2aab73ffc2717e987543defc9b786cc22300aa54b4ba4f3454504e331\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-06T03:34:22Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1206 03:34:15.305330 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1206 03:34:15.307249 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1592846546/tls.crt::/tmp/serving-cert-1592846546/tls.key\\\\\\\"\\\\nI1206 03:34:21.814037 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1206 03:34:21.816756 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1206 03:34:21.816780 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1206 03:34:21.816806 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1206 03:34:21.816813 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1206 03:34:21.831301 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1206 03:34:21.831311 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1206 03:34:21.831324 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 03:34:21.831330 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 03:34:21.831335 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1206 03:34:21.831339 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1206 03:34:21.831342 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1206 03:34:21.831346 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1206 03:34:21.911269 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:04Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d40e0e7446f8416bf9da94fc53c01be937f9a51a11eb101157d2ddd4597720d2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:04Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2cba89f2f1bed96d59f183628d900d9b72c91eb472e273c8bed0babcab485910\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2cba89f2f1bed96d59f183628d900d9b72c91eb472e273c8bed0babcab485910\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:01Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:31Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:31 crc kubenswrapper[4980]: I1206 03:34:31.724654 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6ed68f66-1cf6-465a-8f66-76d5af6a44d8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://725e84f6bdfb722a8ce69158d03ea4c66a9f087642d55cb9651e003c81be0c8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://71f7b492ea07cdf7dc65fc48184ec1e64914eb1811bbfdddc0fb14c8ad6cb296\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b543aa4ee03d5a14909b67d876c5675c02e1df2a89a3c9ac8d3fb587fea53d8f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://32af3427e872a70cb3c2357116f04cabc8d3d9b69e7306b2e0de2ae22f9469d4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:01Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:31Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:31 crc kubenswrapper[4980]: I1206 03:34:31.735826 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:22Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:31Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:31 crc kubenswrapper[4980]: I1206 03:34:31.748184 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://334ea6a273fd7d268906c9f4a6dbf6f4df4fcdbf38f3d7d248cb7d6cf8537794\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ec44dcd5e4bf0a0466bb586ebd61259364dbbfe22499f4cbff854e3bd153082\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:31Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:31 crc kubenswrapper[4980]: I1206 03:34:31.754287 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:34:31 crc kubenswrapper[4980]: I1206 03:34:31.754323 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:34:31 crc kubenswrapper[4980]: I1206 03:34:31.754332 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:34:31 crc kubenswrapper[4980]: I1206 03:34:31.754345 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:34:31 crc kubenswrapper[4980]: I1206 03:34:31.754355 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:34:31Z","lastTransitionTime":"2025-12-06T03:34:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:34:31 crc kubenswrapper[4980]: I1206 03:34:31.762022 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:22Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:31Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:31 crc kubenswrapper[4980]: I1206 03:34:31.773276 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-sv5sz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"873977aa-4eb0-4526-bc0b-757ad2dea2a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01f0cc855c822a71ea8ef6a7ce6e25747ff9956691735973f935c4a28672a7a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lg7tx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:23Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-sv5sz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:31Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:31 crc kubenswrapper[4980]: I1206 03:34:31.792376 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5w4fr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5af0e768-3c9c-4401-ab56-57516cd3170e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:23Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:23Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfzrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfzrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfzrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfzrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfzrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfzrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfzrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfzrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d876df04b5ab0ba07a5cbcafbe4648403c5014c59eb7d40adee2657c3a255473\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d876df04b5ab0ba07a5cbcafbe4648403c5014c59eb7d40adee2657c3a255473\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfzrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:23Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-5w4fr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:31Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:31 crc kubenswrapper[4980]: I1206 03:34:31.811049 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cab608b3-a02f-475b-9aae-9b8f5d878de9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f7fcc44b42bcbb55509048e3a566bb557333c4a83e33d7b0fba8eb514317909\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c22e7dd8886c2dfbb742900ed86ec7b0c2edf9a1363e74f6666961d928f06c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://473e792b2d0a9644854ee0581c6120fc9c989bc342899ef72b1d056080fd4694\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ee8e936dfc39361d6f99460910591ca8c901a20f0d274bbb6aaebfa8a00ee06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://77bc12768b018625ba11a6bfd66688f926c4b7f3bc1d0e41c6551a18b485ecf3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://27cf07cd96893900da4afecd0d56fd9269bc60c1d23246b4685fcd8756357c19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://27cf07cd96893900da4afecd0d56fd9269bc60c1d23246b4685fcd8756357c19\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c49feb24566a8284378670fa0390710d38c237acc94ffb654f90f3fa4626e4c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c49feb24566a8284378670fa0390710d38c237acc94ffb654f90f3fa4626e4c8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:03Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://bb74d1c3d752a379fa394c450217badb45511fbcafad62f7b618131cab94b29e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bb74d1c3d752a379fa394c450217badb45511fbcafad62f7b618131cab94b29e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:01Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:31Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:31 crc kubenswrapper[4980]: I1206 03:34:31.824385 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://502c727d6baa31f400bee14c34b19abbe38af873f28f6a82a43547431755d981\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:31Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:31 crc kubenswrapper[4980]: I1206 03:34:31.839176 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-dx44q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5bac22a-512c-49d0-90db-47de6aaaefd0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:23Z\\\",\\\"message\\\":\\\"containers with incomplete status: [routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:23Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:23Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f7mdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3b5dab476c7f227a6e4a2eae8469432667fb9fad345d9419ba59feab9fd385e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c3b5dab476c7f227a6e4a2eae8469432667fb9fad345d9419ba59feab9fd385e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f7mdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e921499231cbdf26b8918c98f9ad8c322fd7848865caeef048470de3e2c0efba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e921499231cbdf26b8918c98f9ad8c322fd7848865caeef048470de3e2c0efba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f7mdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7cbfe7d8cdb89916c238296592fab45b9e92e2c585baa6d67230decd1c84c5bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7cbfe7d8cdb89916c238296592fab45b9e92e2c585baa6d67230decd1c84c5bb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f7mdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f7mdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f7mdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f7mdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:23Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-dx44q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:31Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:31 crc kubenswrapper[4980]: I1206 03:34:31.850427 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-r5zfb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"320f44d1-a671-4a91-b328-a8b0fdd8f23a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://92393ecaaef8cfcfd01837671f8cc89803dd570b637b4a373b40b7f487788826\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zhm5t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52e8db644463794041978ced2c0591c13e0ada8e5f1a78a94a0a665218a37b5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zhm5t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:23Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-r5zfb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:31Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:31 crc kubenswrapper[4980]: I1206 03:34:31.857282 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:34:31 crc kubenswrapper[4980]: I1206 03:34:31.857322 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:34:31 crc kubenswrapper[4980]: I1206 03:34:31.857332 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:34:31 crc kubenswrapper[4980]: I1206 03:34:31.857348 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:34:31 crc kubenswrapper[4980]: I1206 03:34:31.857361 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:34:31Z","lastTransitionTime":"2025-12-06T03:34:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:34:31 crc kubenswrapper[4980]: I1206 03:34:31.861579 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-cvzg8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f1cde706-268d-4d26-818f-2f42bfb37b5f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://86f9a77acae360740be827eb0d1e1516e8a49cb9b197b6f5f9468ac70a101158\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rk86l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:29Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-cvzg8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:31Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:31 crc kubenswrapper[4980]: I1206 03:34:31.961785 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:34:31 crc kubenswrapper[4980]: I1206 03:34:31.961872 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:34:31 crc kubenswrapper[4980]: I1206 03:34:31.961897 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:34:31 crc kubenswrapper[4980]: I1206 03:34:31.961931 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:34:31 crc kubenswrapper[4980]: I1206 03:34:31.961957 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:34:31Z","lastTransitionTime":"2025-12-06T03:34:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:34:32 crc kubenswrapper[4980]: I1206 03:34:32.065244 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:34:32 crc kubenswrapper[4980]: I1206 03:34:32.065322 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:34:32 crc kubenswrapper[4980]: I1206 03:34:32.065344 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:34:32 crc kubenswrapper[4980]: I1206 03:34:32.065375 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:34:32 crc kubenswrapper[4980]: I1206 03:34:32.065396 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:34:32Z","lastTransitionTime":"2025-12-06T03:34:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:34:32 crc kubenswrapper[4980]: I1206 03:34:32.169195 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:34:32 crc kubenswrapper[4980]: I1206 03:34:32.169253 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:34:32 crc kubenswrapper[4980]: I1206 03:34:32.169280 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:34:32 crc kubenswrapper[4980]: I1206 03:34:32.169306 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:34:32 crc kubenswrapper[4980]: I1206 03:34:32.169323 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:34:32Z","lastTransitionTime":"2025-12-06T03:34:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:34:32 crc kubenswrapper[4980]: I1206 03:34:32.183898 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 03:34:32 crc kubenswrapper[4980]: I1206 03:34:32.183921 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 03:34:32 crc kubenswrapper[4980]: E1206 03:34:32.184047 4980 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 03:34:32 crc kubenswrapper[4980]: E1206 03:34:32.184163 4980 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 03:34:32 crc kubenswrapper[4980]: I1206 03:34:32.272204 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:34:32 crc kubenswrapper[4980]: I1206 03:34:32.272590 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:34:32 crc kubenswrapper[4980]: I1206 03:34:32.272604 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:34:32 crc kubenswrapper[4980]: I1206 03:34:32.272623 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:34:32 crc kubenswrapper[4980]: I1206 03:34:32.272641 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:34:32Z","lastTransitionTime":"2025-12-06T03:34:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:34:32 crc kubenswrapper[4980]: I1206 03:34:32.375543 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:34:32 crc kubenswrapper[4980]: I1206 03:34:32.375625 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:34:32 crc kubenswrapper[4980]: I1206 03:34:32.375642 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:34:32 crc kubenswrapper[4980]: I1206 03:34:32.375663 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:34:32 crc kubenswrapper[4980]: I1206 03:34:32.375684 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:34:32Z","lastTransitionTime":"2025-12-06T03:34:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:34:32 crc kubenswrapper[4980]: I1206 03:34:32.481463 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:34:32 crc kubenswrapper[4980]: I1206 03:34:32.481528 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:34:32 crc kubenswrapper[4980]: I1206 03:34:32.481541 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:34:32 crc kubenswrapper[4980]: I1206 03:34:32.481563 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:34:32 crc kubenswrapper[4980]: I1206 03:34:32.481577 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:34:32Z","lastTransitionTime":"2025-12-06T03:34:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:34:32 crc kubenswrapper[4980]: I1206 03:34:32.585640 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:34:32 crc kubenswrapper[4980]: I1206 03:34:32.585714 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:34:32 crc kubenswrapper[4980]: I1206 03:34:32.585743 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:34:32 crc kubenswrapper[4980]: I1206 03:34:32.585763 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:34:32 crc kubenswrapper[4980]: I1206 03:34:32.585775 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:34:32Z","lastTransitionTime":"2025-12-06T03:34:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:34:32 crc kubenswrapper[4980]: I1206 03:34:32.670733 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5w4fr" event={"ID":"5af0e768-3c9c-4401-ab56-57516cd3170e","Type":"ContainerStarted","Data":"a8d3b4a48fbd8d6d81e7de4c07d3da18b7bfaa209461e10b38daf67000c1f699"} Dec 06 03:34:32 crc kubenswrapper[4980]: I1206 03:34:32.672617 4980 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-5w4fr" Dec 06 03:34:32 crc kubenswrapper[4980]: I1206 03:34:32.672649 4980 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-5w4fr" Dec 06 03:34:32 crc kubenswrapper[4980]: I1206 03:34:32.672699 4980 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-5w4fr" Dec 06 03:34:32 crc kubenswrapper[4980]: I1206 03:34:32.678561 4980 generic.go:334] "Generic (PLEG): container finished" podID="e5bac22a-512c-49d0-90db-47de6aaaefd0" containerID="7f45cbf23b6b0d208cefc4b493f901bab9d71ccf2f1456507d54fb1f30c4fd48" exitCode=0 Dec 06 03:34:32 crc kubenswrapper[4980]: I1206 03:34:32.678560 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-dx44q" event={"ID":"e5bac22a-512c-49d0-90db-47de6aaaefd0","Type":"ContainerDied","Data":"7f45cbf23b6b0d208cefc4b493f901bab9d71ccf2f1456507d54fb1f30c4fd48"} Dec 06 03:34:32 crc kubenswrapper[4980]: I1206 03:34:32.689659 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7f958150-549d-4c80-8a7e-23c39e4f9500\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c28b6895b14e73f4f3eb7a8ce93076c05092ba9b58ea0e3e879c65bcfa227a21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://76cff8a2d652df38b04eefc053832baf4cbb5db778ff5e525b18c2825141b118\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://206f6f19fe3197f5d6798cc86c5473bf2decb02a2860ee7ffb5efb5f60952e8c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bafaa9670b05671fb57b658c2960bf3ee32c84e52f86a3a3bbab01d5a38ff0c4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://773a0bc2aab73ffc2717e987543defc9b786cc22300aa54b4ba4f3454504e331\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-06T03:34:22Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1206 03:34:15.305330 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1206 03:34:15.307249 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1592846546/tls.crt::/tmp/serving-cert-1592846546/tls.key\\\\\\\"\\\\nI1206 03:34:21.814037 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1206 03:34:21.816756 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1206 03:34:21.816780 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1206 03:34:21.816806 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1206 03:34:21.816813 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1206 03:34:21.831301 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1206 03:34:21.831311 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1206 03:34:21.831324 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 03:34:21.831330 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 03:34:21.831335 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1206 03:34:21.831339 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1206 03:34:21.831342 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1206 03:34:21.831346 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1206 03:34:21.911269 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:04Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d40e0e7446f8416bf9da94fc53c01be937f9a51a11eb101157d2ddd4597720d2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:04Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2cba89f2f1bed96d59f183628d900d9b72c91eb472e273c8bed0babcab485910\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2cba89f2f1bed96d59f183628d900d9b72c91eb472e273c8bed0babcab485910\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:01Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:32Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:32 crc kubenswrapper[4980]: I1206 03:34:32.690250 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:34:32 crc kubenswrapper[4980]: I1206 03:34:32.690270 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:34:32 crc kubenswrapper[4980]: I1206 03:34:32.690278 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:34:32 crc kubenswrapper[4980]: I1206 03:34:32.690292 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:34:32 crc kubenswrapper[4980]: I1206 03:34:32.690301 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:34:32Z","lastTransitionTime":"2025-12-06T03:34:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:34:32 crc kubenswrapper[4980]: I1206 03:34:32.736267 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6ed68f66-1cf6-465a-8f66-76d5af6a44d8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://725e84f6bdfb722a8ce69158d03ea4c66a9f087642d55cb9651e003c81be0c8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://71f7b492ea07cdf7dc65fc48184ec1e64914eb1811bbfdddc0fb14c8ad6cb296\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b543aa4ee03d5a14909b67d876c5675c02e1df2a89a3c9ac8d3fb587fea53d8f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://32af3427e872a70cb3c2357116f04cabc8d3d9b69e7306b2e0de2ae22f9469d4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:01Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:32Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:32 crc kubenswrapper[4980]: I1206 03:34:32.746635 4980 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-5w4fr" Dec 06 03:34:32 crc kubenswrapper[4980]: I1206 03:34:32.746902 4980 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-5w4fr" Dec 06 03:34:32 crc kubenswrapper[4980]: I1206 03:34:32.748052 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:22Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:32Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:32 crc kubenswrapper[4980]: I1206 03:34:32.757181 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-sv5sz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"873977aa-4eb0-4526-bc0b-757ad2dea2a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01f0cc855c822a71ea8ef6a7ce6e25747ff9956691735973f935c4a28672a7a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lg7tx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:23Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-sv5sz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:32Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:32 crc kubenswrapper[4980]: I1206 03:34:32.772541 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5w4fr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5af0e768-3c9c-4401-ab56-57516cd3170e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:23Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:23Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://768b8e1fe1a00a8467d744da5ad3f61006b0a3612c96f8e4d0ee6fcb93195f9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfzrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6342903b318317ee8d84c2f5d467cb25c811a789be0d6ab0e0cba37b5611e16d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfzrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d620685574db7594132ae24a91ec0a93dce20a9a289540deebb39650b8a1816\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfzrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c67ec8f9c1764fcdf215d52822c2d4f0437dc7f056f05d708767121f6c1f3515\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfzrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f35beae5339eeaf26494ed678e3c1bd02349eae2a4b4615325b17923f3b653b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfzrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a940ec6ee523313728f7e8ae934b7e0b0cff82771e3e64723309db1fae013c3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfzrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a8d3b4a48fbd8d6d81e7de4c07d3da18b7bfaa209461e10b38daf67000c1f699\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfzrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe261f4322b85702181d7473ab2a9a4d2d4b869d586c3091f811b12a81a63aae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfzrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d876df04b5ab0ba07a5cbcafbe4648403c5014c59eb7d40adee2657c3a255473\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d876df04b5ab0ba07a5cbcafbe4648403c5014c59eb7d40adee2657c3a255473\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfzrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:23Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-5w4fr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:32Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:32 crc kubenswrapper[4980]: I1206 03:34:32.783655 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:22Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:32Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:32 crc kubenswrapper[4980]: I1206 03:34:32.792260 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:34:32 crc kubenswrapper[4980]: I1206 03:34:32.792310 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:34:32 crc kubenswrapper[4980]: I1206 03:34:32.792324 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:34:32 crc kubenswrapper[4980]: I1206 03:34:32.792341 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:34:32 crc kubenswrapper[4980]: I1206 03:34:32.792351 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:34:32Z","lastTransitionTime":"2025-12-06T03:34:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:34:32 crc kubenswrapper[4980]: I1206 03:34:32.794752 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://334ea6a273fd7d268906c9f4a6dbf6f4df4fcdbf38f3d7d248cb7d6cf8537794\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ec44dcd5e4bf0a0466bb586ebd61259364dbbfe22499f4cbff854e3bd153082\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:32Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:32 crc kubenswrapper[4980]: I1206 03:34:32.807267 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-r5zfb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"320f44d1-a671-4a91-b328-a8b0fdd8f23a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://92393ecaaef8cfcfd01837671f8cc89803dd570b637b4a373b40b7f487788826\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zhm5t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52e8db644463794041978ced2c0591c13e0ada8e5f1a78a94a0a665218a37b5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zhm5t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:23Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-r5zfb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:32Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:32 crc kubenswrapper[4980]: I1206 03:34:32.816895 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-cvzg8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f1cde706-268d-4d26-818f-2f42bfb37b5f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://86f9a77acae360740be827eb0d1e1516e8a49cb9b197b6f5f9468ac70a101158\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rk86l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:29Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-cvzg8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:32Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:32 crc kubenswrapper[4980]: I1206 03:34:32.834831 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cab608b3-a02f-475b-9aae-9b8f5d878de9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f7fcc44b42bcbb55509048e3a566bb557333c4a83e33d7b0fba8eb514317909\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c22e7dd8886c2dfbb742900ed86ec7b0c2edf9a1363e74f6666961d928f06c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://473e792b2d0a9644854ee0581c6120fc9c989bc342899ef72b1d056080fd4694\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ee8e936dfc39361d6f99460910591ca8c901a20f0d274bbb6aaebfa8a00ee06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://77bc12768b018625ba11a6bfd66688f926c4b7f3bc1d0e41c6551a18b485ecf3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://27cf07cd96893900da4afecd0d56fd9269bc60c1d23246b4685fcd8756357c19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://27cf07cd96893900da4afecd0d56fd9269bc60c1d23246b4685fcd8756357c19\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c49feb24566a8284378670fa0390710d38c237acc94ffb654f90f3fa4626e4c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c49feb24566a8284378670fa0390710d38c237acc94ffb654f90f3fa4626e4c8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:03Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://bb74d1c3d752a379fa394c450217badb45511fbcafad62f7b618131cab94b29e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bb74d1c3d752a379fa394c450217badb45511fbcafad62f7b618131cab94b29e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:01Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:32Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:32 crc kubenswrapper[4980]: I1206 03:34:32.846245 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://502c727d6baa31f400bee14c34b19abbe38af873f28f6a82a43547431755d981\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:32Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:32 crc kubenswrapper[4980]: I1206 03:34:32.864963 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-dx44q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5bac22a-512c-49d0-90db-47de6aaaefd0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:23Z\\\",\\\"message\\\":\\\"containers with incomplete status: [routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:23Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:23Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f7mdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3b5dab476c7f227a6e4a2eae8469432667fb9fad345d9419ba59feab9fd385e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c3b5dab476c7f227a6e4a2eae8469432667fb9fad345d9419ba59feab9fd385e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f7mdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e921499231cbdf26b8918c98f9ad8c322fd7848865caeef048470de3e2c0efba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e921499231cbdf26b8918c98f9ad8c322fd7848865caeef048470de3e2c0efba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f7mdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7cbfe7d8cdb89916c238296592fab45b9e92e2c585baa6d67230decd1c84c5bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7cbfe7d8cdb89916c238296592fab45b9e92e2c585baa6d67230decd1c84c5bb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f7mdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f7mdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f7mdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f7mdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:23Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-dx44q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:32Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:32 crc kubenswrapper[4980]: I1206 03:34:32.881317 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:32Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:32 crc kubenswrapper[4980]: I1206 03:34:32.894456 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4ef01e7d9809fd9cfc46e3a58b07159cdc6f5a3c210d1c9efc49bca27b30858\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:32Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:32 crc kubenswrapper[4980]: I1206 03:34:32.895505 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:34:32 crc kubenswrapper[4980]: I1206 03:34:32.895569 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:34:32 crc kubenswrapper[4980]: I1206 03:34:32.895580 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:34:32 crc kubenswrapper[4980]: I1206 03:34:32.895947 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:34:32 crc kubenswrapper[4980]: I1206 03:34:32.895978 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:34:32Z","lastTransitionTime":"2025-12-06T03:34:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:34:32 crc kubenswrapper[4980]: I1206 03:34:32.905881 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-kq4z8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f96c1f6c-55e8-4d86-bff3-5381581b7022\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://799498681afa7653408bc23b1e82d75c16e3c809d2ee7f322780e20c537eb710\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bdx7g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:23Z\\\"}}\" for pod \"openshift-multus\"/\"multus-kq4z8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:32Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:32 crc kubenswrapper[4980]: I1206 03:34:32.924289 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:22Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:32Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:32 crc kubenswrapper[4980]: I1206 03:34:32.936369 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://334ea6a273fd7d268906c9f4a6dbf6f4df4fcdbf38f3d7d248cb7d6cf8537794\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ec44dcd5e4bf0a0466bb586ebd61259364dbbfe22499f4cbff854e3bd153082\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:32Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:32 crc kubenswrapper[4980]: I1206 03:34:32.959010 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:22Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:32Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:32 crc kubenswrapper[4980]: I1206 03:34:32.972319 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-sv5sz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"873977aa-4eb0-4526-bc0b-757ad2dea2a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01f0cc855c822a71ea8ef6a7ce6e25747ff9956691735973f935c4a28672a7a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lg7tx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:23Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-sv5sz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:32Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:32 crc kubenswrapper[4980]: I1206 03:34:32.994748 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5w4fr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5af0e768-3c9c-4401-ab56-57516cd3170e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:23Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:23Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://768b8e1fe1a00a8467d744da5ad3f61006b0a3612c96f8e4d0ee6fcb93195f9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfzrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6342903b318317ee8d84c2f5d467cb25c811a789be0d6ab0e0cba37b5611e16d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfzrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d620685574db7594132ae24a91ec0a93dce20a9a289540deebb39650b8a1816\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfzrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c67ec8f9c1764fcdf215d52822c2d4f0437dc7f056f05d708767121f6c1f3515\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfzrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f35beae5339eeaf26494ed678e3c1bd02349eae2a4b4615325b17923f3b653b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfzrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a940ec6ee523313728f7e8ae934b7e0b0cff82771e3e64723309db1fae013c3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfzrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a8d3b4a48fbd8d6d81e7de4c07d3da18b7bfaa209461e10b38daf67000c1f699\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfzrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe261f4322b85702181d7473ab2a9a4d2d4b869d586c3091f811b12a81a63aae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfzrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d876df04b5ab0ba07a5cbcafbe4648403c5014c59eb7d40adee2657c3a255473\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d876df04b5ab0ba07a5cbcafbe4648403c5014c59eb7d40adee2657c3a255473\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfzrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:23Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-5w4fr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:32Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:33 crc kubenswrapper[4980]: I1206 03:34:33.000153 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:34:33 crc kubenswrapper[4980]: I1206 03:34:33.000193 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:34:33 crc kubenswrapper[4980]: I1206 03:34:33.000202 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:34:33 crc kubenswrapper[4980]: I1206 03:34:33.000218 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:34:33 crc kubenswrapper[4980]: I1206 03:34:33.000229 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:34:33Z","lastTransitionTime":"2025-12-06T03:34:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:34:33 crc kubenswrapper[4980]: I1206 03:34:33.022309 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cab608b3-a02f-475b-9aae-9b8f5d878de9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f7fcc44b42bcbb55509048e3a566bb557333c4a83e33d7b0fba8eb514317909\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c22e7dd8886c2dfbb742900ed86ec7b0c2edf9a1363e74f6666961d928f06c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://473e792b2d0a9644854ee0581c6120fc9c989bc342899ef72b1d056080fd4694\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ee8e936dfc39361d6f99460910591ca8c901a20f0d274bbb6aaebfa8a00ee06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://77bc12768b018625ba11a6bfd66688f926c4b7f3bc1d0e41c6551a18b485ecf3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://27cf07cd96893900da4afecd0d56fd9269bc60c1d23246b4685fcd8756357c19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://27cf07cd96893900da4afecd0d56fd9269bc60c1d23246b4685fcd8756357c19\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c49feb24566a8284378670fa0390710d38c237acc94ffb654f90f3fa4626e4c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c49feb24566a8284378670fa0390710d38c237acc94ffb654f90f3fa4626e4c8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:03Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://bb74d1c3d752a379fa394c450217badb45511fbcafad62f7b618131cab94b29e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bb74d1c3d752a379fa394c450217badb45511fbcafad62f7b618131cab94b29e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:01Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:33Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:33 crc kubenswrapper[4980]: I1206 03:34:33.034222 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://502c727d6baa31f400bee14c34b19abbe38af873f28f6a82a43547431755d981\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:33Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:33 crc kubenswrapper[4980]: I1206 03:34:33.047974 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-dx44q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5bac22a-512c-49d0-90db-47de6aaaefd0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:23Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:23Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:23Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f7mdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3b5dab476c7f227a6e4a2eae8469432667fb9fad345d9419ba59feab9fd385e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c3b5dab476c7f227a6e4a2eae8469432667fb9fad345d9419ba59feab9fd385e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f7mdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e921499231cbdf26b8918c98f9ad8c322fd7848865caeef048470de3e2c0efba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e921499231cbdf26b8918c98f9ad8c322fd7848865caeef048470de3e2c0efba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f7mdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7cbfe7d8cdb89916c238296592fab45b9e92e2c585baa6d67230decd1c84c5bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7cbfe7d8cdb89916c238296592fab45b9e92e2c585baa6d67230decd1c84c5bb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f7mdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f45cbf23b6b0d208cefc4b493f901bab9d71ccf2f1456507d54fb1f30c4fd48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7f45cbf23b6b0d208cefc4b493f901bab9d71ccf2f1456507d54fb1f30c4fd48\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f7mdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f7mdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f7mdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:23Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-dx44q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:33Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:33 crc kubenswrapper[4980]: I1206 03:34:33.060324 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-r5zfb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"320f44d1-a671-4a91-b328-a8b0fdd8f23a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://92393ecaaef8cfcfd01837671f8cc89803dd570b637b4a373b40b7f487788826\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zhm5t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52e8db644463794041978ced2c0591c13e0ada8e5f1a78a94a0a665218a37b5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zhm5t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:23Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-r5zfb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:33Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:33 crc kubenswrapper[4980]: I1206 03:34:33.072217 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-cvzg8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f1cde706-268d-4d26-818f-2f42bfb37b5f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://86f9a77acae360740be827eb0d1e1516e8a49cb9b197b6f5f9468ac70a101158\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rk86l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:29Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-cvzg8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:33Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:33 crc kubenswrapper[4980]: I1206 03:34:33.085412 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:33Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:33 crc kubenswrapper[4980]: I1206 03:34:33.102397 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4ef01e7d9809fd9cfc46e3a58b07159cdc6f5a3c210d1c9efc49bca27b30858\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:33Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:33 crc kubenswrapper[4980]: I1206 03:34:33.114847 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-kq4z8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f96c1f6c-55e8-4d86-bff3-5381581b7022\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://799498681afa7653408bc23b1e82d75c16e3c809d2ee7f322780e20c537eb710\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bdx7g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:23Z\\\"}}\" for pod \"openshift-multus\"/\"multus-kq4z8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:33Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:33 crc kubenswrapper[4980]: I1206 03:34:33.129862 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7f958150-549d-4c80-8a7e-23c39e4f9500\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c28b6895b14e73f4f3eb7a8ce93076c05092ba9b58ea0e3e879c65bcfa227a21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://76cff8a2d652df38b04eefc053832baf4cbb5db778ff5e525b18c2825141b118\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://206f6f19fe3197f5d6798cc86c5473bf2decb02a2860ee7ffb5efb5f60952e8c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bafaa9670b05671fb57b658c2960bf3ee32c84e52f86a3a3bbab01d5a38ff0c4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://773a0bc2aab73ffc2717e987543defc9b786cc22300aa54b4ba4f3454504e331\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-06T03:34:22Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1206 03:34:15.305330 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1206 03:34:15.307249 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1592846546/tls.crt::/tmp/serving-cert-1592846546/tls.key\\\\\\\"\\\\nI1206 03:34:21.814037 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1206 03:34:21.816756 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1206 03:34:21.816780 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1206 03:34:21.816806 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1206 03:34:21.816813 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1206 03:34:21.831301 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1206 03:34:21.831311 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1206 03:34:21.831324 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 03:34:21.831330 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 03:34:21.831335 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1206 03:34:21.831339 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1206 03:34:21.831342 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1206 03:34:21.831346 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1206 03:34:21.911269 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:04Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d40e0e7446f8416bf9da94fc53c01be937f9a51a11eb101157d2ddd4597720d2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:04Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2cba89f2f1bed96d59f183628d900d9b72c91eb472e273c8bed0babcab485910\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2cba89f2f1bed96d59f183628d900d9b72c91eb472e273c8bed0babcab485910\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:01Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:33Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:33 crc kubenswrapper[4980]: I1206 03:34:33.139467 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:34:33 crc kubenswrapper[4980]: I1206 03:34:33.139504 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:34:33 crc kubenswrapper[4980]: I1206 03:34:33.139530 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:34:33 crc kubenswrapper[4980]: I1206 03:34:33.139544 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:34:33 crc kubenswrapper[4980]: I1206 03:34:33.139554 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:34:33Z","lastTransitionTime":"2025-12-06T03:34:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:34:33 crc kubenswrapper[4980]: I1206 03:34:33.144697 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6ed68f66-1cf6-465a-8f66-76d5af6a44d8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://725e84f6bdfb722a8ce69158d03ea4c66a9f087642d55cb9651e003c81be0c8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://71f7b492ea07cdf7dc65fc48184ec1e64914eb1811bbfdddc0fb14c8ad6cb296\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b543aa4ee03d5a14909b67d876c5675c02e1df2a89a3c9ac8d3fb587fea53d8f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://32af3427e872a70cb3c2357116f04cabc8d3d9b69e7306b2e0de2ae22f9469d4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:01Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:33Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:33 crc kubenswrapper[4980]: I1206 03:34:33.186827 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 03:34:33 crc kubenswrapper[4980]: E1206 03:34:33.187001 4980 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 03:34:33 crc kubenswrapper[4980]: I1206 03:34:33.244426 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:34:33 crc kubenswrapper[4980]: I1206 03:34:33.244481 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:34:33 crc kubenswrapper[4980]: I1206 03:34:33.244499 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:34:33 crc kubenswrapper[4980]: I1206 03:34:33.244744 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:34:33 crc kubenswrapper[4980]: I1206 03:34:33.244782 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:34:33Z","lastTransitionTime":"2025-12-06T03:34:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:34:33 crc kubenswrapper[4980]: I1206 03:34:33.347345 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:34:33 crc kubenswrapper[4980]: I1206 03:34:33.347382 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:34:33 crc kubenswrapper[4980]: I1206 03:34:33.347391 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:34:33 crc kubenswrapper[4980]: I1206 03:34:33.347413 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:34:33 crc kubenswrapper[4980]: I1206 03:34:33.347423 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:34:33Z","lastTransitionTime":"2025-12-06T03:34:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:34:33 crc kubenswrapper[4980]: I1206 03:34:33.451218 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:34:33 crc kubenswrapper[4980]: I1206 03:34:33.451308 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:34:33 crc kubenswrapper[4980]: I1206 03:34:33.451327 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:34:33 crc kubenswrapper[4980]: I1206 03:34:33.451354 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:34:33 crc kubenswrapper[4980]: I1206 03:34:33.451369 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:34:33Z","lastTransitionTime":"2025-12-06T03:34:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:34:33 crc kubenswrapper[4980]: I1206 03:34:33.554761 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:34:33 crc kubenswrapper[4980]: I1206 03:34:33.554813 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:34:33 crc kubenswrapper[4980]: I1206 03:34:33.554828 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:34:33 crc kubenswrapper[4980]: I1206 03:34:33.554851 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:34:33 crc kubenswrapper[4980]: I1206 03:34:33.554869 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:34:33Z","lastTransitionTime":"2025-12-06T03:34:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:34:33 crc kubenswrapper[4980]: I1206 03:34:33.658136 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:34:33 crc kubenswrapper[4980]: I1206 03:34:33.658204 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:34:33 crc kubenswrapper[4980]: I1206 03:34:33.658221 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:34:33 crc kubenswrapper[4980]: I1206 03:34:33.658264 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:34:33 crc kubenswrapper[4980]: I1206 03:34:33.658282 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:34:33Z","lastTransitionTime":"2025-12-06T03:34:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:34:33 crc kubenswrapper[4980]: I1206 03:34:33.687808 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-dx44q" event={"ID":"e5bac22a-512c-49d0-90db-47de6aaaefd0","Type":"ContainerStarted","Data":"344ce4b145ff021fcaf163803d8ad8cc6e9565bf7f0992e2653c43db663d8411"} Dec 06 03:34:33 crc kubenswrapper[4980]: I1206 03:34:33.711855 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cab608b3-a02f-475b-9aae-9b8f5d878de9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f7fcc44b42bcbb55509048e3a566bb557333c4a83e33d7b0fba8eb514317909\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c22e7dd8886c2dfbb742900ed86ec7b0c2edf9a1363e74f6666961d928f06c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://473e792b2d0a9644854ee0581c6120fc9c989bc342899ef72b1d056080fd4694\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ee8e936dfc39361d6f99460910591ca8c901a20f0d274bbb6aaebfa8a00ee06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://77bc12768b018625ba11a6bfd66688f926c4b7f3bc1d0e41c6551a18b485ecf3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://27cf07cd96893900da4afecd0d56fd9269bc60c1d23246b4685fcd8756357c19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://27cf07cd96893900da4afecd0d56fd9269bc60c1d23246b4685fcd8756357c19\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c49feb24566a8284378670fa0390710d38c237acc94ffb654f90f3fa4626e4c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c49feb24566a8284378670fa0390710d38c237acc94ffb654f90f3fa4626e4c8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:03Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://bb74d1c3d752a379fa394c450217badb45511fbcafad62f7b618131cab94b29e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bb74d1c3d752a379fa394c450217badb45511fbcafad62f7b618131cab94b29e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:01Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:33Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:33 crc kubenswrapper[4980]: I1206 03:34:33.727832 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://502c727d6baa31f400bee14c34b19abbe38af873f28f6a82a43547431755d981\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:33Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:33 crc kubenswrapper[4980]: I1206 03:34:33.750970 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-dx44q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5bac22a-512c-49d0-90db-47de6aaaefd0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:23Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:23Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:23Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f7mdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3b5dab476c7f227a6e4a2eae8469432667fb9fad345d9419ba59feab9fd385e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c3b5dab476c7f227a6e4a2eae8469432667fb9fad345d9419ba59feab9fd385e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f7mdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e921499231cbdf26b8918c98f9ad8c322fd7848865caeef048470de3e2c0efba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e921499231cbdf26b8918c98f9ad8c322fd7848865caeef048470de3e2c0efba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f7mdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7cbfe7d8cdb89916c238296592fab45b9e92e2c585baa6d67230decd1c84c5bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7cbfe7d8cdb89916c238296592fab45b9e92e2c585baa6d67230decd1c84c5bb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f7mdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f45cbf23b6b0d208cefc4b493f901bab9d71ccf2f1456507d54fb1f30c4fd48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7f45cbf23b6b0d208cefc4b493f901bab9d71ccf2f1456507d54fb1f30c4fd48\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f7mdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://344ce4b145ff021fcaf163803d8ad8cc6e9565bf7f0992e2653c43db663d8411\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f7mdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f7mdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:23Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-dx44q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:33Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:33 crc kubenswrapper[4980]: I1206 03:34:33.762045 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:34:33 crc kubenswrapper[4980]: I1206 03:34:33.762086 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:34:33 crc kubenswrapper[4980]: I1206 03:34:33.762096 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:34:33 crc kubenswrapper[4980]: I1206 03:34:33.762114 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:34:33 crc kubenswrapper[4980]: I1206 03:34:33.762128 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:34:33Z","lastTransitionTime":"2025-12-06T03:34:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:34:33 crc kubenswrapper[4980]: I1206 03:34:33.766648 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-r5zfb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"320f44d1-a671-4a91-b328-a8b0fdd8f23a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://92393ecaaef8cfcfd01837671f8cc89803dd570b637b4a373b40b7f487788826\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zhm5t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52e8db644463794041978ced2c0591c13e0ada8e5f1a78a94a0a665218a37b5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zhm5t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:23Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-r5zfb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:33Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:33 crc kubenswrapper[4980]: I1206 03:34:33.779721 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-cvzg8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f1cde706-268d-4d26-818f-2f42bfb37b5f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://86f9a77acae360740be827eb0d1e1516e8a49cb9b197b6f5f9468ac70a101158\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rk86l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:29Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-cvzg8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:33Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:33 crc kubenswrapper[4980]: I1206 03:34:33.794750 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:33Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:33 crc kubenswrapper[4980]: I1206 03:34:33.813593 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4ef01e7d9809fd9cfc46e3a58b07159cdc6f5a3c210d1c9efc49bca27b30858\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:33Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:33 crc kubenswrapper[4980]: I1206 03:34:33.829345 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-kq4z8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f96c1f6c-55e8-4d86-bff3-5381581b7022\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://799498681afa7653408bc23b1e82d75c16e3c809d2ee7f322780e20c537eb710\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bdx7g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:23Z\\\"}}\" for pod \"openshift-multus\"/\"multus-kq4z8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:33Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:33 crc kubenswrapper[4980]: I1206 03:34:33.845165 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7f958150-549d-4c80-8a7e-23c39e4f9500\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c28b6895b14e73f4f3eb7a8ce93076c05092ba9b58ea0e3e879c65bcfa227a21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://76cff8a2d652df38b04eefc053832baf4cbb5db778ff5e525b18c2825141b118\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://206f6f19fe3197f5d6798cc86c5473bf2decb02a2860ee7ffb5efb5f60952e8c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bafaa9670b05671fb57b658c2960bf3ee32c84e52f86a3a3bbab01d5a38ff0c4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://773a0bc2aab73ffc2717e987543defc9b786cc22300aa54b4ba4f3454504e331\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-06T03:34:22Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1206 03:34:15.305330 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1206 03:34:15.307249 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1592846546/tls.crt::/tmp/serving-cert-1592846546/tls.key\\\\\\\"\\\\nI1206 03:34:21.814037 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1206 03:34:21.816756 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1206 03:34:21.816780 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1206 03:34:21.816806 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1206 03:34:21.816813 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1206 03:34:21.831301 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1206 03:34:21.831311 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1206 03:34:21.831324 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 03:34:21.831330 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 03:34:21.831335 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1206 03:34:21.831339 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1206 03:34:21.831342 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1206 03:34:21.831346 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1206 03:34:21.911269 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:04Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d40e0e7446f8416bf9da94fc53c01be937f9a51a11eb101157d2ddd4597720d2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:04Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2cba89f2f1bed96d59f183628d900d9b72c91eb472e273c8bed0babcab485910\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2cba89f2f1bed96d59f183628d900d9b72c91eb472e273c8bed0babcab485910\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:01Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:33Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:33 crc kubenswrapper[4980]: I1206 03:34:33.859977 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6ed68f66-1cf6-465a-8f66-76d5af6a44d8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://725e84f6bdfb722a8ce69158d03ea4c66a9f087642d55cb9651e003c81be0c8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://71f7b492ea07cdf7dc65fc48184ec1e64914eb1811bbfdddc0fb14c8ad6cb296\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b543aa4ee03d5a14909b67d876c5675c02e1df2a89a3c9ac8d3fb587fea53d8f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://32af3427e872a70cb3c2357116f04cabc8d3d9b69e7306b2e0de2ae22f9469d4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:01Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:33Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:33 crc kubenswrapper[4980]: I1206 03:34:33.883495 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:34:33 crc kubenswrapper[4980]: I1206 03:34:33.883573 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:34:33 crc kubenswrapper[4980]: I1206 03:34:33.883584 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:34:33 crc kubenswrapper[4980]: I1206 03:34:33.883606 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:34:33 crc kubenswrapper[4980]: I1206 03:34:33.883618 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:34:33Z","lastTransitionTime":"2025-12-06T03:34:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:34:33 crc kubenswrapper[4980]: I1206 03:34:33.884544 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5w4fr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5af0e768-3c9c-4401-ab56-57516cd3170e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:23Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:23Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://768b8e1fe1a00a8467d744da5ad3f61006b0a3612c96f8e4d0ee6fcb93195f9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfzrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6342903b318317ee8d84c2f5d467cb25c811a789be0d6ab0e0cba37b5611e16d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfzrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d620685574db7594132ae24a91ec0a93dce20a9a289540deebb39650b8a1816\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfzrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c67ec8f9c1764fcdf215d52822c2d4f0437dc7f056f05d708767121f6c1f3515\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfzrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f35beae5339eeaf26494ed678e3c1bd02349eae2a4b4615325b17923f3b653b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfzrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a940ec6ee523313728f7e8ae934b7e0b0cff82771e3e64723309db1fae013c3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfzrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a8d3b4a48fbd8d6d81e7de4c07d3da18b7bfaa209461e10b38daf67000c1f699\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfzrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe261f4322b85702181d7473ab2a9a4d2d4b869d586c3091f811b12a81a63aae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfzrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d876df04b5ab0ba07a5cbcafbe4648403c5014c59eb7d40adee2657c3a255473\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d876df04b5ab0ba07a5cbcafbe4648403c5014c59eb7d40adee2657c3a255473\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfzrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:23Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-5w4fr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:33Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:33 crc kubenswrapper[4980]: I1206 03:34:33.898441 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:22Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:33Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:33 crc kubenswrapper[4980]: I1206 03:34:33.910786 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://334ea6a273fd7d268906c9f4a6dbf6f4df4fcdbf38f3d7d248cb7d6cf8537794\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ec44dcd5e4bf0a0466bb586ebd61259364dbbfe22499f4cbff854e3bd153082\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:33Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:33 crc kubenswrapper[4980]: I1206 03:34:33.926056 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:22Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:33Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:33 crc kubenswrapper[4980]: I1206 03:34:33.937856 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-sv5sz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"873977aa-4eb0-4526-bc0b-757ad2dea2a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01f0cc855c822a71ea8ef6a7ce6e25747ff9956691735973f935c4a28672a7a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lg7tx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:23Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-sv5sz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:33Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:33 crc kubenswrapper[4980]: I1206 03:34:33.986367 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:34:33 crc kubenswrapper[4980]: I1206 03:34:33.986409 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:34:33 crc kubenswrapper[4980]: I1206 03:34:33.986420 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:34:33 crc kubenswrapper[4980]: I1206 03:34:33.986435 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:34:33 crc kubenswrapper[4980]: I1206 03:34:33.986446 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:34:33Z","lastTransitionTime":"2025-12-06T03:34:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:34:34 crc kubenswrapper[4980]: I1206 03:34:34.090014 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:34:34 crc kubenswrapper[4980]: I1206 03:34:34.090066 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:34:34 crc kubenswrapper[4980]: I1206 03:34:34.090277 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:34:34 crc kubenswrapper[4980]: I1206 03:34:34.090307 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:34:34 crc kubenswrapper[4980]: I1206 03:34:34.090325 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:34:34Z","lastTransitionTime":"2025-12-06T03:34:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:34:34 crc kubenswrapper[4980]: I1206 03:34:34.184125 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 03:34:34 crc kubenswrapper[4980]: E1206 03:34:34.184275 4980 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 03:34:34 crc kubenswrapper[4980]: I1206 03:34:34.184124 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 03:34:34 crc kubenswrapper[4980]: E1206 03:34:34.184656 4980 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 03:34:34 crc kubenswrapper[4980]: I1206 03:34:34.193209 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:34:34 crc kubenswrapper[4980]: I1206 03:34:34.193244 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:34:34 crc kubenswrapper[4980]: I1206 03:34:34.193252 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:34:34 crc kubenswrapper[4980]: I1206 03:34:34.193266 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:34:34 crc kubenswrapper[4980]: I1206 03:34:34.193276 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:34:34Z","lastTransitionTime":"2025-12-06T03:34:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:34:34 crc kubenswrapper[4980]: I1206 03:34:34.295834 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:34:34 crc kubenswrapper[4980]: I1206 03:34:34.296108 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:34:34 crc kubenswrapper[4980]: I1206 03:34:34.296178 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:34:34 crc kubenswrapper[4980]: I1206 03:34:34.296272 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:34:34 crc kubenswrapper[4980]: I1206 03:34:34.296339 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:34:34Z","lastTransitionTime":"2025-12-06T03:34:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:34:34 crc kubenswrapper[4980]: I1206 03:34:34.399797 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:34:34 crc kubenswrapper[4980]: I1206 03:34:34.399826 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:34:34 crc kubenswrapper[4980]: I1206 03:34:34.399836 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:34:34 crc kubenswrapper[4980]: I1206 03:34:34.399850 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:34:34 crc kubenswrapper[4980]: I1206 03:34:34.399858 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:34:34Z","lastTransitionTime":"2025-12-06T03:34:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:34:34 crc kubenswrapper[4980]: I1206 03:34:34.502660 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:34:34 crc kubenswrapper[4980]: I1206 03:34:34.502706 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:34:34 crc kubenswrapper[4980]: I1206 03:34:34.502717 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:34:34 crc kubenswrapper[4980]: I1206 03:34:34.502732 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:34:34 crc kubenswrapper[4980]: I1206 03:34:34.502742 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:34:34Z","lastTransitionTime":"2025-12-06T03:34:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:34:34 crc kubenswrapper[4980]: I1206 03:34:34.605148 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:34:34 crc kubenswrapper[4980]: I1206 03:34:34.605198 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:34:34 crc kubenswrapper[4980]: I1206 03:34:34.605208 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:34:34 crc kubenswrapper[4980]: I1206 03:34:34.605222 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:34:34 crc kubenswrapper[4980]: I1206 03:34:34.605230 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:34:34Z","lastTransitionTime":"2025-12-06T03:34:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:34:34 crc kubenswrapper[4980]: I1206 03:34:34.708406 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:34:34 crc kubenswrapper[4980]: I1206 03:34:34.708458 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:34:34 crc kubenswrapper[4980]: I1206 03:34:34.708467 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:34:34 crc kubenswrapper[4980]: I1206 03:34:34.708485 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:34:34 crc kubenswrapper[4980]: I1206 03:34:34.708502 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:34:34Z","lastTransitionTime":"2025-12-06T03:34:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:34:34 crc kubenswrapper[4980]: I1206 03:34:34.817299 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:34:34 crc kubenswrapper[4980]: I1206 03:34:34.817363 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:34:34 crc kubenswrapper[4980]: I1206 03:34:34.817376 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:34:34 crc kubenswrapper[4980]: I1206 03:34:34.817405 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:34:34 crc kubenswrapper[4980]: I1206 03:34:34.817420 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:34:34Z","lastTransitionTime":"2025-12-06T03:34:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:34:34 crc kubenswrapper[4980]: I1206 03:34:34.920927 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:34:34 crc kubenswrapper[4980]: I1206 03:34:34.921437 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:34:34 crc kubenswrapper[4980]: I1206 03:34:34.921704 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:34:34 crc kubenswrapper[4980]: I1206 03:34:34.921929 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:34:34 crc kubenswrapper[4980]: I1206 03:34:34.922113 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:34:34Z","lastTransitionTime":"2025-12-06T03:34:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:34:35 crc kubenswrapper[4980]: I1206 03:34:35.025559 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:34:35 crc kubenswrapper[4980]: I1206 03:34:35.026048 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:34:35 crc kubenswrapper[4980]: I1206 03:34:35.026139 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:34:35 crc kubenswrapper[4980]: I1206 03:34:35.026259 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:34:35 crc kubenswrapper[4980]: I1206 03:34:35.026344 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:34:35Z","lastTransitionTime":"2025-12-06T03:34:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:34:35 crc kubenswrapper[4980]: I1206 03:34:35.129481 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:34:35 crc kubenswrapper[4980]: I1206 03:34:35.129549 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:34:35 crc kubenswrapper[4980]: I1206 03:34:35.129568 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:34:35 crc kubenswrapper[4980]: I1206 03:34:35.129592 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:34:35 crc kubenswrapper[4980]: I1206 03:34:35.129607 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:34:35Z","lastTransitionTime":"2025-12-06T03:34:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:34:35 crc kubenswrapper[4980]: I1206 03:34:35.184495 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 03:34:35 crc kubenswrapper[4980]: E1206 03:34:35.184724 4980 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 03:34:35 crc kubenswrapper[4980]: I1206 03:34:35.232703 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:34:35 crc kubenswrapper[4980]: I1206 03:34:35.232756 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:34:35 crc kubenswrapper[4980]: I1206 03:34:35.232765 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:34:35 crc kubenswrapper[4980]: I1206 03:34:35.232816 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:34:35 crc kubenswrapper[4980]: I1206 03:34:35.232828 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:34:35Z","lastTransitionTime":"2025-12-06T03:34:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:34:35 crc kubenswrapper[4980]: I1206 03:34:35.336223 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:34:35 crc kubenswrapper[4980]: I1206 03:34:35.336263 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:34:35 crc kubenswrapper[4980]: I1206 03:34:35.336274 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:34:35 crc kubenswrapper[4980]: I1206 03:34:35.336291 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:34:35 crc kubenswrapper[4980]: I1206 03:34:35.336302 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:34:35Z","lastTransitionTime":"2025-12-06T03:34:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:34:35 crc kubenswrapper[4980]: I1206 03:34:35.440482 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:34:35 crc kubenswrapper[4980]: I1206 03:34:35.440623 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:34:35 crc kubenswrapper[4980]: I1206 03:34:35.440661 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:34:35 crc kubenswrapper[4980]: I1206 03:34:35.440697 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:34:35 crc kubenswrapper[4980]: I1206 03:34:35.440720 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:34:35Z","lastTransitionTime":"2025-12-06T03:34:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:34:35 crc kubenswrapper[4980]: I1206 03:34:35.544603 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:34:35 crc kubenswrapper[4980]: I1206 03:34:35.544656 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:34:35 crc kubenswrapper[4980]: I1206 03:34:35.544671 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:34:35 crc kubenswrapper[4980]: I1206 03:34:35.544695 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:34:35 crc kubenswrapper[4980]: I1206 03:34:35.544713 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:34:35Z","lastTransitionTime":"2025-12-06T03:34:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:34:35 crc kubenswrapper[4980]: I1206 03:34:35.665453 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:34:35 crc kubenswrapper[4980]: I1206 03:34:35.665490 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:34:35 crc kubenswrapper[4980]: I1206 03:34:35.665501 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:34:35 crc kubenswrapper[4980]: I1206 03:34:35.665532 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:34:35 crc kubenswrapper[4980]: I1206 03:34:35.665541 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:34:35Z","lastTransitionTime":"2025-12-06T03:34:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:34:35 crc kubenswrapper[4980]: I1206 03:34:35.695931 4980 generic.go:334] "Generic (PLEG): container finished" podID="e5bac22a-512c-49d0-90db-47de6aaaefd0" containerID="344ce4b145ff021fcaf163803d8ad8cc6e9565bf7f0992e2653c43db663d8411" exitCode=0 Dec 06 03:34:35 crc kubenswrapper[4980]: I1206 03:34:35.695979 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-dx44q" event={"ID":"e5bac22a-512c-49d0-90db-47de6aaaefd0","Type":"ContainerDied","Data":"344ce4b145ff021fcaf163803d8ad8cc6e9565bf7f0992e2653c43db663d8411"} Dec 06 03:34:35 crc kubenswrapper[4980]: I1206 03:34:35.719382 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-sv5sz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"873977aa-4eb0-4526-bc0b-757ad2dea2a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01f0cc855c822a71ea8ef6a7ce6e25747ff9956691735973f935c4a28672a7a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lg7tx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:23Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-sv5sz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:35Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:35 crc kubenswrapper[4980]: I1206 03:34:35.753084 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5w4fr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5af0e768-3c9c-4401-ab56-57516cd3170e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:23Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:23Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://768b8e1fe1a00a8467d744da5ad3f61006b0a3612c96f8e4d0ee6fcb93195f9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfzrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6342903b318317ee8d84c2f5d467cb25c811a789be0d6ab0e0cba37b5611e16d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfzrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d620685574db7594132ae24a91ec0a93dce20a9a289540deebb39650b8a1816\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfzrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c67ec8f9c1764fcdf215d52822c2d4f0437dc7f056f05d708767121f6c1f3515\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfzrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f35beae5339eeaf26494ed678e3c1bd02349eae2a4b4615325b17923f3b653b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfzrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a940ec6ee523313728f7e8ae934b7e0b0cff82771e3e64723309db1fae013c3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfzrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a8d3b4a48fbd8d6d81e7de4c07d3da18b7bfaa209461e10b38daf67000c1f699\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfzrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe261f4322b85702181d7473ab2a9a4d2d4b869d586c3091f811b12a81a63aae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfzrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d876df04b5ab0ba07a5cbcafbe4648403c5014c59eb7d40adee2657c3a255473\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d876df04b5ab0ba07a5cbcafbe4648403c5014c59eb7d40adee2657c3a255473\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfzrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:23Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-5w4fr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:35Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:35 crc kubenswrapper[4980]: I1206 03:34:35.766836 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:22Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:35Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:35 crc kubenswrapper[4980]: I1206 03:34:35.767812 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:34:35 crc kubenswrapper[4980]: I1206 03:34:35.767862 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:34:35 crc kubenswrapper[4980]: I1206 03:34:35.767874 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:34:35 crc kubenswrapper[4980]: I1206 03:34:35.767896 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:34:35 crc kubenswrapper[4980]: I1206 03:34:35.767908 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:34:35Z","lastTransitionTime":"2025-12-06T03:34:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:34:35 crc kubenswrapper[4980]: I1206 03:34:35.781058 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://334ea6a273fd7d268906c9f4a6dbf6f4df4fcdbf38f3d7d248cb7d6cf8537794\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ec44dcd5e4bf0a0466bb586ebd61259364dbbfe22499f4cbff854e3bd153082\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:35Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:35 crc kubenswrapper[4980]: I1206 03:34:35.796235 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:22Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:35Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:35 crc kubenswrapper[4980]: I1206 03:34:35.806334 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-cvzg8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f1cde706-268d-4d26-818f-2f42bfb37b5f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://86f9a77acae360740be827eb0d1e1516e8a49cb9b197b6f5f9468ac70a101158\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rk86l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:29Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-cvzg8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:35Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:35 crc kubenswrapper[4980]: I1206 03:34:35.825436 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cab608b3-a02f-475b-9aae-9b8f5d878de9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f7fcc44b42bcbb55509048e3a566bb557333c4a83e33d7b0fba8eb514317909\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c22e7dd8886c2dfbb742900ed86ec7b0c2edf9a1363e74f6666961d928f06c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://473e792b2d0a9644854ee0581c6120fc9c989bc342899ef72b1d056080fd4694\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ee8e936dfc39361d6f99460910591ca8c901a20f0d274bbb6aaebfa8a00ee06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://77bc12768b018625ba11a6bfd66688f926c4b7f3bc1d0e41c6551a18b485ecf3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://27cf07cd96893900da4afecd0d56fd9269bc60c1d23246b4685fcd8756357c19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://27cf07cd96893900da4afecd0d56fd9269bc60c1d23246b4685fcd8756357c19\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c49feb24566a8284378670fa0390710d38c237acc94ffb654f90f3fa4626e4c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c49feb24566a8284378670fa0390710d38c237acc94ffb654f90f3fa4626e4c8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:03Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://bb74d1c3d752a379fa394c450217badb45511fbcafad62f7b618131cab94b29e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bb74d1c3d752a379fa394c450217badb45511fbcafad62f7b618131cab94b29e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:01Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:35Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:35 crc kubenswrapper[4980]: I1206 03:34:35.838777 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://502c727d6baa31f400bee14c34b19abbe38af873f28f6a82a43547431755d981\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:35Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:35 crc kubenswrapper[4980]: I1206 03:34:35.871219 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:34:35 crc kubenswrapper[4980]: I1206 03:34:35.871253 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:34:35 crc kubenswrapper[4980]: I1206 03:34:35.871262 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:34:35 crc kubenswrapper[4980]: I1206 03:34:35.871277 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:34:35 crc kubenswrapper[4980]: I1206 03:34:35.871286 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:34:35Z","lastTransitionTime":"2025-12-06T03:34:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:34:35 crc kubenswrapper[4980]: I1206 03:34:35.894431 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-dx44q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5bac22a-512c-49d0-90db-47de6aaaefd0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:23Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:23Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:23Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f7mdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3b5dab476c7f227a6e4a2eae8469432667fb9fad345d9419ba59feab9fd385e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c3b5dab476c7f227a6e4a2eae8469432667fb9fad345d9419ba59feab9fd385e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f7mdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e921499231cbdf26b8918c98f9ad8c322fd7848865caeef048470de3e2c0efba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e921499231cbdf26b8918c98f9ad8c322fd7848865caeef048470de3e2c0efba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f7mdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7cbfe7d8cdb89916c238296592fab45b9e92e2c585baa6d67230decd1c84c5bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7cbfe7d8cdb89916c238296592fab45b9e92e2c585baa6d67230decd1c84c5bb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f7mdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f45cbf23b6b0d208cefc4b493f901bab9d71ccf2f1456507d54fb1f30c4fd48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7f45cbf23b6b0d208cefc4b493f901bab9d71ccf2f1456507d54fb1f30c4fd48\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f7mdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://344ce4b145ff021fcaf163803d8ad8cc6e9565bf7f0992e2653c43db663d8411\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://344ce4b145ff021fcaf163803d8ad8cc6e9565bf7f0992e2653c43db663d8411\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f7mdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f7mdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:23Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-dx44q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:35Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:35 crc kubenswrapper[4980]: I1206 03:34:35.911409 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-r5zfb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"320f44d1-a671-4a91-b328-a8b0fdd8f23a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://92393ecaaef8cfcfd01837671f8cc89803dd570b637b4a373b40b7f487788826\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zhm5t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52e8db644463794041978ced2c0591c13e0ada8e5f1a78a94a0a665218a37b5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zhm5t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:23Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-r5zfb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:35Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:35 crc kubenswrapper[4980]: I1206 03:34:35.924051 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:35Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:35 crc kubenswrapper[4980]: I1206 03:34:35.937913 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4ef01e7d9809fd9cfc46e3a58b07159cdc6f5a3c210d1c9efc49bca27b30858\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:35Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:35 crc kubenswrapper[4980]: I1206 03:34:35.949643 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-kq4z8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f96c1f6c-55e8-4d86-bff3-5381581b7022\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://799498681afa7653408bc23b1e82d75c16e3c809d2ee7f322780e20c537eb710\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bdx7g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:23Z\\\"}}\" for pod \"openshift-multus\"/\"multus-kq4z8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:35Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:35 crc kubenswrapper[4980]: I1206 03:34:35.963565 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7f958150-549d-4c80-8a7e-23c39e4f9500\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c28b6895b14e73f4f3eb7a8ce93076c05092ba9b58ea0e3e879c65bcfa227a21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://76cff8a2d652df38b04eefc053832baf4cbb5db778ff5e525b18c2825141b118\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://206f6f19fe3197f5d6798cc86c5473bf2decb02a2860ee7ffb5efb5f60952e8c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bafaa9670b05671fb57b658c2960bf3ee32c84e52f86a3a3bbab01d5a38ff0c4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://773a0bc2aab73ffc2717e987543defc9b786cc22300aa54b4ba4f3454504e331\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-06T03:34:22Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1206 03:34:15.305330 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1206 03:34:15.307249 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1592846546/tls.crt::/tmp/serving-cert-1592846546/tls.key\\\\\\\"\\\\nI1206 03:34:21.814037 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1206 03:34:21.816756 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1206 03:34:21.816780 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1206 03:34:21.816806 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1206 03:34:21.816813 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1206 03:34:21.831301 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1206 03:34:21.831311 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1206 03:34:21.831324 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 03:34:21.831330 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 03:34:21.831335 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1206 03:34:21.831339 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1206 03:34:21.831342 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1206 03:34:21.831346 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1206 03:34:21.911269 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:04Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d40e0e7446f8416bf9da94fc53c01be937f9a51a11eb101157d2ddd4597720d2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:04Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2cba89f2f1bed96d59f183628d900d9b72c91eb472e273c8bed0babcab485910\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2cba89f2f1bed96d59f183628d900d9b72c91eb472e273c8bed0babcab485910\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:01Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:35Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:35 crc kubenswrapper[4980]: I1206 03:34:35.973661 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:34:35 crc kubenswrapper[4980]: I1206 03:34:35.973713 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:34:35 crc kubenswrapper[4980]: I1206 03:34:35.973721 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:34:35 crc kubenswrapper[4980]: I1206 03:34:35.973735 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:34:35 crc kubenswrapper[4980]: I1206 03:34:35.973749 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:34:35Z","lastTransitionTime":"2025-12-06T03:34:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:34:35 crc kubenswrapper[4980]: I1206 03:34:35.975403 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6ed68f66-1cf6-465a-8f66-76d5af6a44d8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://725e84f6bdfb722a8ce69158d03ea4c66a9f087642d55cb9651e003c81be0c8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://71f7b492ea07cdf7dc65fc48184ec1e64914eb1811bbfdddc0fb14c8ad6cb296\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b543aa4ee03d5a14909b67d876c5675c02e1df2a89a3c9ac8d3fb587fea53d8f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://32af3427e872a70cb3c2357116f04cabc8d3d9b69e7306b2e0de2ae22f9469d4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:01Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:35Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:36 crc kubenswrapper[4980]: I1206 03:34:36.076770 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:34:36 crc kubenswrapper[4980]: I1206 03:34:36.076807 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:34:36 crc kubenswrapper[4980]: I1206 03:34:36.076817 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:34:36 crc kubenswrapper[4980]: I1206 03:34:36.076832 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:34:36 crc kubenswrapper[4980]: I1206 03:34:36.076841 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:34:36Z","lastTransitionTime":"2025-12-06T03:34:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:34:36 crc kubenswrapper[4980]: I1206 03:34:36.179481 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:34:36 crc kubenswrapper[4980]: I1206 03:34:36.179569 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:34:36 crc kubenswrapper[4980]: I1206 03:34:36.179589 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:34:36 crc kubenswrapper[4980]: I1206 03:34:36.179612 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:34:36 crc kubenswrapper[4980]: I1206 03:34:36.179627 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:34:36Z","lastTransitionTime":"2025-12-06T03:34:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:34:36 crc kubenswrapper[4980]: I1206 03:34:36.184011 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 03:34:36 crc kubenswrapper[4980]: I1206 03:34:36.184042 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 03:34:36 crc kubenswrapper[4980]: E1206 03:34:36.184165 4980 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 03:34:36 crc kubenswrapper[4980]: E1206 03:34:36.184277 4980 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 03:34:36 crc kubenswrapper[4980]: I1206 03:34:36.286995 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:34:36 crc kubenswrapper[4980]: I1206 03:34:36.287044 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:34:36 crc kubenswrapper[4980]: I1206 03:34:36.287054 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:34:36 crc kubenswrapper[4980]: I1206 03:34:36.287071 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:34:36 crc kubenswrapper[4980]: I1206 03:34:36.287081 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:34:36Z","lastTransitionTime":"2025-12-06T03:34:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:34:36 crc kubenswrapper[4980]: I1206 03:34:36.390602 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:34:36 crc kubenswrapper[4980]: I1206 03:34:36.390644 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:34:36 crc kubenswrapper[4980]: I1206 03:34:36.390656 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:34:36 crc kubenswrapper[4980]: I1206 03:34:36.390674 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:34:36 crc kubenswrapper[4980]: I1206 03:34:36.390685 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:34:36Z","lastTransitionTime":"2025-12-06T03:34:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:34:36 crc kubenswrapper[4980]: I1206 03:34:36.493465 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:34:36 crc kubenswrapper[4980]: I1206 03:34:36.493552 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:34:36 crc kubenswrapper[4980]: I1206 03:34:36.493571 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:34:36 crc kubenswrapper[4980]: I1206 03:34:36.493598 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:34:36 crc kubenswrapper[4980]: I1206 03:34:36.493614 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:34:36Z","lastTransitionTime":"2025-12-06T03:34:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:34:36 crc kubenswrapper[4980]: I1206 03:34:36.597506 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:34:36 crc kubenswrapper[4980]: I1206 03:34:36.597571 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:34:36 crc kubenswrapper[4980]: I1206 03:34:36.597584 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:34:36 crc kubenswrapper[4980]: I1206 03:34:36.597603 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:34:36 crc kubenswrapper[4980]: I1206 03:34:36.597621 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:34:36Z","lastTransitionTime":"2025-12-06T03:34:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:34:36 crc kubenswrapper[4980]: I1206 03:34:36.699983 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:34:36 crc kubenswrapper[4980]: I1206 03:34:36.700023 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:34:36 crc kubenswrapper[4980]: I1206 03:34:36.700035 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:34:36 crc kubenswrapper[4980]: I1206 03:34:36.700052 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:34:36 crc kubenswrapper[4980]: I1206 03:34:36.700065 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:34:36Z","lastTransitionTime":"2025-12-06T03:34:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:34:36 crc kubenswrapper[4980]: I1206 03:34:36.803607 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:34:36 crc kubenswrapper[4980]: I1206 03:34:36.803670 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:34:36 crc kubenswrapper[4980]: I1206 03:34:36.803688 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:34:36 crc kubenswrapper[4980]: I1206 03:34:36.803713 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:34:36 crc kubenswrapper[4980]: I1206 03:34:36.803734 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:34:36Z","lastTransitionTime":"2025-12-06T03:34:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:34:36 crc kubenswrapper[4980]: I1206 03:34:36.907764 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:34:36 crc kubenswrapper[4980]: I1206 03:34:36.908061 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:34:36 crc kubenswrapper[4980]: I1206 03:34:36.908210 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:34:36 crc kubenswrapper[4980]: I1206 03:34:36.908329 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:34:36 crc kubenswrapper[4980]: I1206 03:34:36.908435 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:34:36Z","lastTransitionTime":"2025-12-06T03:34:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:34:37 crc kubenswrapper[4980]: I1206 03:34:37.011279 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:34:37 crc kubenswrapper[4980]: I1206 03:34:37.011313 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:34:37 crc kubenswrapper[4980]: I1206 03:34:37.011322 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:34:37 crc kubenswrapper[4980]: I1206 03:34:37.011337 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:34:37 crc kubenswrapper[4980]: I1206 03:34:37.011347 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:34:37Z","lastTransitionTime":"2025-12-06T03:34:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:34:37 crc kubenswrapper[4980]: I1206 03:34:37.114715 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:34:37 crc kubenswrapper[4980]: I1206 03:34:37.114774 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:34:37 crc kubenswrapper[4980]: I1206 03:34:37.114791 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:34:37 crc kubenswrapper[4980]: I1206 03:34:37.114817 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:34:37 crc kubenswrapper[4980]: I1206 03:34:37.114834 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:34:37Z","lastTransitionTime":"2025-12-06T03:34:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:34:37 crc kubenswrapper[4980]: I1206 03:34:37.183852 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 03:34:37 crc kubenswrapper[4980]: E1206 03:34:37.184043 4980 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 03:34:37 crc kubenswrapper[4980]: I1206 03:34:37.218087 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:34:37 crc kubenswrapper[4980]: I1206 03:34:37.218132 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:34:37 crc kubenswrapper[4980]: I1206 03:34:37.218144 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:34:37 crc kubenswrapper[4980]: I1206 03:34:37.218162 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:34:37 crc kubenswrapper[4980]: I1206 03:34:37.218173 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:34:37Z","lastTransitionTime":"2025-12-06T03:34:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:34:37 crc kubenswrapper[4980]: I1206 03:34:37.320991 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:34:37 crc kubenswrapper[4980]: I1206 03:34:37.321030 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:34:37 crc kubenswrapper[4980]: I1206 03:34:37.321039 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:34:37 crc kubenswrapper[4980]: I1206 03:34:37.321055 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:34:37 crc kubenswrapper[4980]: I1206 03:34:37.321066 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:34:37Z","lastTransitionTime":"2025-12-06T03:34:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:34:37 crc kubenswrapper[4980]: I1206 03:34:37.351308 4980 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-6kxfq"] Dec 06 03:34:37 crc kubenswrapper[4980]: I1206 03:34:37.351939 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-6kxfq" Dec 06 03:34:37 crc kubenswrapper[4980]: I1206 03:34:37.353752 4980 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Dec 06 03:34:37 crc kubenswrapper[4980]: I1206 03:34:37.355908 4980 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Dec 06 03:34:37 crc kubenswrapper[4980]: I1206 03:34:37.375410 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cab608b3-a02f-475b-9aae-9b8f5d878de9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f7fcc44b42bcbb55509048e3a566bb557333c4a83e33d7b0fba8eb514317909\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c22e7dd8886c2dfbb742900ed86ec7b0c2edf9a1363e74f6666961d928f06c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://473e792b2d0a9644854ee0581c6120fc9c989bc342899ef72b1d056080fd4694\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ee8e936dfc39361d6f99460910591ca8c901a20f0d274bbb6aaebfa8a00ee06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://77bc12768b018625ba11a6bfd66688f926c4b7f3bc1d0e41c6551a18b485ecf3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://27cf07cd96893900da4afecd0d56fd9269bc60c1d23246b4685fcd8756357c19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://27cf07cd96893900da4afecd0d56fd9269bc60c1d23246b4685fcd8756357c19\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c49feb24566a8284378670fa0390710d38c237acc94ffb654f90f3fa4626e4c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c49feb24566a8284378670fa0390710d38c237acc94ffb654f90f3fa4626e4c8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:03Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://bb74d1c3d752a379fa394c450217badb45511fbcafad62f7b618131cab94b29e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bb74d1c3d752a379fa394c450217badb45511fbcafad62f7b618131cab94b29e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:01Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:37Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:37 crc kubenswrapper[4980]: I1206 03:34:37.388203 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://502c727d6baa31f400bee14c34b19abbe38af873f28f6a82a43547431755d981\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:37Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:37 crc kubenswrapper[4980]: I1206 03:34:37.404674 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-dx44q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5bac22a-512c-49d0-90db-47de6aaaefd0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:23Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:23Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:23Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f7mdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3b5dab476c7f227a6e4a2eae8469432667fb9fad345d9419ba59feab9fd385e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c3b5dab476c7f227a6e4a2eae8469432667fb9fad345d9419ba59feab9fd385e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f7mdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e921499231cbdf26b8918c98f9ad8c322fd7848865caeef048470de3e2c0efba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e921499231cbdf26b8918c98f9ad8c322fd7848865caeef048470de3e2c0efba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f7mdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7cbfe7d8cdb89916c238296592fab45b9e92e2c585baa6d67230decd1c84c5bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7cbfe7d8cdb89916c238296592fab45b9e92e2c585baa6d67230decd1c84c5bb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f7mdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f45cbf23b6b0d208cefc4b493f901bab9d71ccf2f1456507d54fb1f30c4fd48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7f45cbf23b6b0d208cefc4b493f901bab9d71ccf2f1456507d54fb1f30c4fd48\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f7mdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://344ce4b145ff021fcaf163803d8ad8cc6e9565bf7f0992e2653c43db663d8411\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://344ce4b145ff021fcaf163803d8ad8cc6e9565bf7f0992e2653c43db663d8411\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f7mdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f7mdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:23Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-dx44q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:37Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:37 crc kubenswrapper[4980]: I1206 03:34:37.418158 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-r5zfb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"320f44d1-a671-4a91-b328-a8b0fdd8f23a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://92393ecaaef8cfcfd01837671f8cc89803dd570b637b4a373b40b7f487788826\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zhm5t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52e8db644463794041978ced2c0591c13e0ada8e5f1a78a94a0a665218a37b5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zhm5t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:23Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-r5zfb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:37Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:37 crc kubenswrapper[4980]: I1206 03:34:37.424087 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:34:37 crc kubenswrapper[4980]: I1206 03:34:37.424145 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:34:37 crc kubenswrapper[4980]: I1206 03:34:37.424166 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:34:37 crc kubenswrapper[4980]: I1206 03:34:37.424191 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:34:37 crc kubenswrapper[4980]: I1206 03:34:37.424208 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:34:37Z","lastTransitionTime":"2025-12-06T03:34:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:34:37 crc kubenswrapper[4980]: I1206 03:34:37.428647 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/951b6deb-bcea-4332-8ae8-c23ac7aaef36-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-6kxfq\" (UID: \"951b6deb-bcea-4332-8ae8-c23ac7aaef36\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-6kxfq" Dec 06 03:34:37 crc kubenswrapper[4980]: I1206 03:34:37.428782 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/951b6deb-bcea-4332-8ae8-c23ac7aaef36-env-overrides\") pod \"ovnkube-control-plane-749d76644c-6kxfq\" (UID: \"951b6deb-bcea-4332-8ae8-c23ac7aaef36\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-6kxfq" Dec 06 03:34:37 crc kubenswrapper[4980]: I1206 03:34:37.428835 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/951b6deb-bcea-4332-8ae8-c23ac7aaef36-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-6kxfq\" (UID: \"951b6deb-bcea-4332-8ae8-c23ac7aaef36\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-6kxfq" Dec 06 03:34:37 crc kubenswrapper[4980]: I1206 03:34:37.428873 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b8htf\" (UniqueName: \"kubernetes.io/projected/951b6deb-bcea-4332-8ae8-c23ac7aaef36-kube-api-access-b8htf\") pod \"ovnkube-control-plane-749d76644c-6kxfq\" (UID: \"951b6deb-bcea-4332-8ae8-c23ac7aaef36\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-6kxfq" Dec 06 03:34:37 crc kubenswrapper[4980]: I1206 03:34:37.431343 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-cvzg8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f1cde706-268d-4d26-818f-2f42bfb37b5f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://86f9a77acae360740be827eb0d1e1516e8a49cb9b197b6f5f9468ac70a101158\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rk86l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:29Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-cvzg8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:37Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:37 crc kubenswrapper[4980]: I1206 03:34:37.445443 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:37Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:37 crc kubenswrapper[4980]: I1206 03:34:37.458577 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4ef01e7d9809fd9cfc46e3a58b07159cdc6f5a3c210d1c9efc49bca27b30858\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:37Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:37 crc kubenswrapper[4980]: I1206 03:34:37.470666 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-kq4z8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f96c1f6c-55e8-4d86-bff3-5381581b7022\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://799498681afa7653408bc23b1e82d75c16e3c809d2ee7f322780e20c537eb710\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bdx7g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:23Z\\\"}}\" for pod \"openshift-multus\"/\"multus-kq4z8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:37Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:37 crc kubenswrapper[4980]: I1206 03:34:37.480419 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-6kxfq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"951b6deb-bcea-4332-8ae8-c23ac7aaef36\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:37Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:37Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b8htf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b8htf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:37Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-6kxfq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:37Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:37 crc kubenswrapper[4980]: I1206 03:34:37.496284 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7f958150-549d-4c80-8a7e-23c39e4f9500\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c28b6895b14e73f4f3eb7a8ce93076c05092ba9b58ea0e3e879c65bcfa227a21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://76cff8a2d652df38b04eefc053832baf4cbb5db778ff5e525b18c2825141b118\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://206f6f19fe3197f5d6798cc86c5473bf2decb02a2860ee7ffb5efb5f60952e8c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bafaa9670b05671fb57b658c2960bf3ee32c84e52f86a3a3bbab01d5a38ff0c4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://773a0bc2aab73ffc2717e987543defc9b786cc22300aa54b4ba4f3454504e331\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-06T03:34:22Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1206 03:34:15.305330 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1206 03:34:15.307249 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1592846546/tls.crt::/tmp/serving-cert-1592846546/tls.key\\\\\\\"\\\\nI1206 03:34:21.814037 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1206 03:34:21.816756 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1206 03:34:21.816780 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1206 03:34:21.816806 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1206 03:34:21.816813 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1206 03:34:21.831301 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1206 03:34:21.831311 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1206 03:34:21.831324 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 03:34:21.831330 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 03:34:21.831335 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1206 03:34:21.831339 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1206 03:34:21.831342 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1206 03:34:21.831346 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1206 03:34:21.911269 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:04Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d40e0e7446f8416bf9da94fc53c01be937f9a51a11eb101157d2ddd4597720d2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:04Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2cba89f2f1bed96d59f183628d900d9b72c91eb472e273c8bed0babcab485910\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2cba89f2f1bed96d59f183628d900d9b72c91eb472e273c8bed0babcab485910\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:01Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:37Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:37 crc kubenswrapper[4980]: I1206 03:34:37.511632 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6ed68f66-1cf6-465a-8f66-76d5af6a44d8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://725e84f6bdfb722a8ce69158d03ea4c66a9f087642d55cb9651e003c81be0c8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://71f7b492ea07cdf7dc65fc48184ec1e64914eb1811bbfdddc0fb14c8ad6cb296\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b543aa4ee03d5a14909b67d876c5675c02e1df2a89a3c9ac8d3fb587fea53d8f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://32af3427e872a70cb3c2357116f04cabc8d3d9b69e7306b2e0de2ae22f9469d4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:01Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:37Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:37 crc kubenswrapper[4980]: I1206 03:34:37.526173 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:34:37 crc kubenswrapper[4980]: I1206 03:34:37.526217 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:34:37 crc kubenswrapper[4980]: I1206 03:34:37.526229 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:34:37 crc kubenswrapper[4980]: I1206 03:34:37.526247 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:34:37 crc kubenswrapper[4980]: I1206 03:34:37.526261 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:34:37Z","lastTransitionTime":"2025-12-06T03:34:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:34:37 crc kubenswrapper[4980]: I1206 03:34:37.529638 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/951b6deb-bcea-4332-8ae8-c23ac7aaef36-env-overrides\") pod \"ovnkube-control-plane-749d76644c-6kxfq\" (UID: \"951b6deb-bcea-4332-8ae8-c23ac7aaef36\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-6kxfq" Dec 06 03:34:37 crc kubenswrapper[4980]: I1206 03:34:37.529695 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/951b6deb-bcea-4332-8ae8-c23ac7aaef36-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-6kxfq\" (UID: \"951b6deb-bcea-4332-8ae8-c23ac7aaef36\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-6kxfq" Dec 06 03:34:37 crc kubenswrapper[4980]: I1206 03:34:37.529733 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b8htf\" (UniqueName: \"kubernetes.io/projected/951b6deb-bcea-4332-8ae8-c23ac7aaef36-kube-api-access-b8htf\") pod \"ovnkube-control-plane-749d76644c-6kxfq\" (UID: \"951b6deb-bcea-4332-8ae8-c23ac7aaef36\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-6kxfq" Dec 06 03:34:37 crc kubenswrapper[4980]: I1206 03:34:37.529763 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/951b6deb-bcea-4332-8ae8-c23ac7aaef36-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-6kxfq\" (UID: \"951b6deb-bcea-4332-8ae8-c23ac7aaef36\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-6kxfq" Dec 06 03:34:37 crc kubenswrapper[4980]: I1206 03:34:37.530551 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/951b6deb-bcea-4332-8ae8-c23ac7aaef36-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-6kxfq\" (UID: \"951b6deb-bcea-4332-8ae8-c23ac7aaef36\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-6kxfq" Dec 06 03:34:37 crc kubenswrapper[4980]: I1206 03:34:37.531222 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/951b6deb-bcea-4332-8ae8-c23ac7aaef36-env-overrides\") pod \"ovnkube-control-plane-749d76644c-6kxfq\" (UID: \"951b6deb-bcea-4332-8ae8-c23ac7aaef36\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-6kxfq" Dec 06 03:34:37 crc kubenswrapper[4980]: I1206 03:34:37.536178 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/951b6deb-bcea-4332-8ae8-c23ac7aaef36-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-6kxfq\" (UID: \"951b6deb-bcea-4332-8ae8-c23ac7aaef36\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-6kxfq" Dec 06 03:34:37 crc kubenswrapper[4980]: I1206 03:34:37.536936 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5w4fr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5af0e768-3c9c-4401-ab56-57516cd3170e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:23Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:23Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://768b8e1fe1a00a8467d744da5ad3f61006b0a3612c96f8e4d0ee6fcb93195f9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfzrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6342903b318317ee8d84c2f5d467cb25c811a789be0d6ab0e0cba37b5611e16d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfzrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d620685574db7594132ae24a91ec0a93dce20a9a289540deebb39650b8a1816\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfzrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c67ec8f9c1764fcdf215d52822c2d4f0437dc7f056f05d708767121f6c1f3515\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfzrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f35beae5339eeaf26494ed678e3c1bd02349eae2a4b4615325b17923f3b653b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfzrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a940ec6ee523313728f7e8ae934b7e0b0cff82771e3e64723309db1fae013c3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfzrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a8d3b4a48fbd8d6d81e7de4c07d3da18b7bfaa209461e10b38daf67000c1f699\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfzrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe261f4322b85702181d7473ab2a9a4d2d4b869d586c3091f811b12a81a63aae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfzrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d876df04b5ab0ba07a5cbcafbe4648403c5014c59eb7d40adee2657c3a255473\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d876df04b5ab0ba07a5cbcafbe4648403c5014c59eb7d40adee2657c3a255473\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfzrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:23Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-5w4fr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:37Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:37 crc kubenswrapper[4980]: I1206 03:34:37.552567 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b8htf\" (UniqueName: \"kubernetes.io/projected/951b6deb-bcea-4332-8ae8-c23ac7aaef36-kube-api-access-b8htf\") pod \"ovnkube-control-plane-749d76644c-6kxfq\" (UID: \"951b6deb-bcea-4332-8ae8-c23ac7aaef36\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-6kxfq" Dec 06 03:34:37 crc kubenswrapper[4980]: I1206 03:34:37.566052 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:22Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:37Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:37 crc kubenswrapper[4980]: I1206 03:34:37.567454 4980 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 06 03:34:37 crc kubenswrapper[4980]: I1206 03:34:37.584299 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://334ea6a273fd7d268906c9f4a6dbf6f4df4fcdbf38f3d7d248cb7d6cf8537794\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ec44dcd5e4bf0a0466bb586ebd61259364dbbfe22499f4cbff854e3bd153082\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:37Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:37 crc kubenswrapper[4980]: I1206 03:34:37.596125 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:22Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:37Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:37 crc kubenswrapper[4980]: I1206 03:34:37.605688 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-sv5sz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"873977aa-4eb0-4526-bc0b-757ad2dea2a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01f0cc855c822a71ea8ef6a7ce6e25747ff9956691735973f935c4a28672a7a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lg7tx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:23Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-sv5sz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:37Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:37 crc kubenswrapper[4980]: I1206 03:34:37.615577 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:22Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:37Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:37 crc kubenswrapper[4980]: I1206 03:34:37.628975 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://334ea6a273fd7d268906c9f4a6dbf6f4df4fcdbf38f3d7d248cb7d6cf8537794\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ec44dcd5e4bf0a0466bb586ebd61259364dbbfe22499f4cbff854e3bd153082\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:37Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:37 crc kubenswrapper[4980]: I1206 03:34:37.629217 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:34:37 crc kubenswrapper[4980]: I1206 03:34:37.629282 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:34:37 crc kubenswrapper[4980]: I1206 03:34:37.629293 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:34:37 crc kubenswrapper[4980]: I1206 03:34:37.629310 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:34:37 crc kubenswrapper[4980]: I1206 03:34:37.629325 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:34:37Z","lastTransitionTime":"2025-12-06T03:34:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:34:37 crc kubenswrapper[4980]: I1206 03:34:37.641202 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:22Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:37Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:37 crc kubenswrapper[4980]: I1206 03:34:37.652106 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-sv5sz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"873977aa-4eb0-4526-bc0b-757ad2dea2a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01f0cc855c822a71ea8ef6a7ce6e25747ff9956691735973f935c4a28672a7a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lg7tx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:23Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-sv5sz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:37Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:37 crc kubenswrapper[4980]: I1206 03:34:37.666795 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-6kxfq" Dec 06 03:34:37 crc kubenswrapper[4980]: I1206 03:34:37.672504 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5w4fr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5af0e768-3c9c-4401-ab56-57516cd3170e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:23Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:23Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://768b8e1fe1a00a8467d744da5ad3f61006b0a3612c96f8e4d0ee6fcb93195f9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfzrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6342903b318317ee8d84c2f5d467cb25c811a789be0d6ab0e0cba37b5611e16d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfzrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d620685574db7594132ae24a91ec0a93dce20a9a289540deebb39650b8a1816\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfzrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c67ec8f9c1764fcdf215d52822c2d4f0437dc7f056f05d708767121f6c1f3515\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfzrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f35beae5339eeaf26494ed678e3c1bd02349eae2a4b4615325b17923f3b653b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfzrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a940ec6ee523313728f7e8ae934b7e0b0cff82771e3e64723309db1fae013c3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfzrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a8d3b4a48fbd8d6d81e7de4c07d3da18b7bfaa209461e10b38daf67000c1f699\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfzrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe261f4322b85702181d7473ab2a9a4d2d4b869d586c3091f811b12a81a63aae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfzrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d876df04b5ab0ba07a5cbcafbe4648403c5014c59eb7d40adee2657c3a255473\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d876df04b5ab0ba07a5cbcafbe4648403c5014c59eb7d40adee2657c3a255473\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfzrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:23Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-5w4fr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:37Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:37 crc kubenswrapper[4980]: I1206 03:34:37.692714 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cab608b3-a02f-475b-9aae-9b8f5d878de9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f7fcc44b42bcbb55509048e3a566bb557333c4a83e33d7b0fba8eb514317909\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c22e7dd8886c2dfbb742900ed86ec7b0c2edf9a1363e74f6666961d928f06c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://473e792b2d0a9644854ee0581c6120fc9c989bc342899ef72b1d056080fd4694\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ee8e936dfc39361d6f99460910591ca8c901a20f0d274bbb6aaebfa8a00ee06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://77bc12768b018625ba11a6bfd66688f926c4b7f3bc1d0e41c6551a18b485ecf3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://27cf07cd96893900da4afecd0d56fd9269bc60c1d23246b4685fcd8756357c19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://27cf07cd96893900da4afecd0d56fd9269bc60c1d23246b4685fcd8756357c19\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c49feb24566a8284378670fa0390710d38c237acc94ffb654f90f3fa4626e4c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c49feb24566a8284378670fa0390710d38c237acc94ffb654f90f3fa4626e4c8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:03Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://bb74d1c3d752a379fa394c450217badb45511fbcafad62f7b618131cab94b29e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bb74d1c3d752a379fa394c450217badb45511fbcafad62f7b618131cab94b29e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:01Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:37Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:37 crc kubenswrapper[4980]: I1206 03:34:37.703354 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-6kxfq" event={"ID":"951b6deb-bcea-4332-8ae8-c23ac7aaef36","Type":"ContainerStarted","Data":"62de139cb4980c66d7ac415e062d5c2efa70f61729e6886f61bf6eae22e3a0b2"} Dec 06 03:34:37 crc kubenswrapper[4980]: I1206 03:34:37.705079 4980 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-5w4fr_5af0e768-3c9c-4401-ab56-57516cd3170e/ovnkube-controller/0.log" Dec 06 03:34:37 crc kubenswrapper[4980]: I1206 03:34:37.705264 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://502c727d6baa31f400bee14c34b19abbe38af873f28f6a82a43547431755d981\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:37Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:37 crc kubenswrapper[4980]: I1206 03:34:37.708489 4980 generic.go:334] "Generic (PLEG): container finished" podID="5af0e768-3c9c-4401-ab56-57516cd3170e" containerID="a8d3b4a48fbd8d6d81e7de4c07d3da18b7bfaa209461e10b38daf67000c1f699" exitCode=1 Dec 06 03:34:37 crc kubenswrapper[4980]: I1206 03:34:37.708551 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5w4fr" event={"ID":"5af0e768-3c9c-4401-ab56-57516cd3170e","Type":"ContainerDied","Data":"a8d3b4a48fbd8d6d81e7de4c07d3da18b7bfaa209461e10b38daf67000c1f699"} Dec 06 03:34:37 crc kubenswrapper[4980]: I1206 03:34:37.709188 4980 scope.go:117] "RemoveContainer" containerID="a8d3b4a48fbd8d6d81e7de4c07d3da18b7bfaa209461e10b38daf67000c1f699" Dec 06 03:34:37 crc kubenswrapper[4980]: I1206 03:34:37.713603 4980 generic.go:334] "Generic (PLEG): container finished" podID="e5bac22a-512c-49d0-90db-47de6aaaefd0" containerID="619e56a9937dd6376ab471ba16da30e5d77f443f5bd33610bea5ec85cd0446ef" exitCode=0 Dec 06 03:34:37 crc kubenswrapper[4980]: I1206 03:34:37.713633 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-dx44q" event={"ID":"e5bac22a-512c-49d0-90db-47de6aaaefd0","Type":"ContainerDied","Data":"619e56a9937dd6376ab471ba16da30e5d77f443f5bd33610bea5ec85cd0446ef"} Dec 06 03:34:37 crc kubenswrapper[4980]: I1206 03:34:37.719468 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-dx44q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5bac22a-512c-49d0-90db-47de6aaaefd0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:23Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:23Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:23Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f7mdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3b5dab476c7f227a6e4a2eae8469432667fb9fad345d9419ba59feab9fd385e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c3b5dab476c7f227a6e4a2eae8469432667fb9fad345d9419ba59feab9fd385e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f7mdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e921499231cbdf26b8918c98f9ad8c322fd7848865caeef048470de3e2c0efba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e921499231cbdf26b8918c98f9ad8c322fd7848865caeef048470de3e2c0efba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f7mdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7cbfe7d8cdb89916c238296592fab45b9e92e2c585baa6d67230decd1c84c5bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7cbfe7d8cdb89916c238296592fab45b9e92e2c585baa6d67230decd1c84c5bb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f7mdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f45cbf23b6b0d208cefc4b493f901bab9d71ccf2f1456507d54fb1f30c4fd48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7f45cbf23b6b0d208cefc4b493f901bab9d71ccf2f1456507d54fb1f30c4fd48\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f7mdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://344ce4b145ff021fcaf163803d8ad8cc6e9565bf7f0992e2653c43db663d8411\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://344ce4b145ff021fcaf163803d8ad8cc6e9565bf7f0992e2653c43db663d8411\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f7mdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f7mdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:23Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-dx44q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:37Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:37 crc kubenswrapper[4980]: I1206 03:34:37.731345 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-r5zfb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"320f44d1-a671-4a91-b328-a8b0fdd8f23a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://92393ecaaef8cfcfd01837671f8cc89803dd570b637b4a373b40b7f487788826\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zhm5t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52e8db644463794041978ced2c0591c13e0ada8e5f1a78a94a0a665218a37b5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zhm5t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:23Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-r5zfb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:37Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:37 crc kubenswrapper[4980]: I1206 03:34:37.732853 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:34:37 crc kubenswrapper[4980]: I1206 03:34:37.732873 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:34:37 crc kubenswrapper[4980]: I1206 03:34:37.732882 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:34:37 crc kubenswrapper[4980]: I1206 03:34:37.732895 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:34:37 crc kubenswrapper[4980]: I1206 03:34:37.732903 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:34:37Z","lastTransitionTime":"2025-12-06T03:34:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:34:37 crc kubenswrapper[4980]: I1206 03:34:37.740405 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-cvzg8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f1cde706-268d-4d26-818f-2f42bfb37b5f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://86f9a77acae360740be827eb0d1e1516e8a49cb9b197b6f5f9468ac70a101158\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rk86l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:29Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-cvzg8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:37Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:37 crc kubenswrapper[4980]: I1206 03:34:37.751818 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:37Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:37 crc kubenswrapper[4980]: I1206 03:34:37.768940 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4ef01e7d9809fd9cfc46e3a58b07159cdc6f5a3c210d1c9efc49bca27b30858\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:37Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:37 crc kubenswrapper[4980]: I1206 03:34:37.782333 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-kq4z8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f96c1f6c-55e8-4d86-bff3-5381581b7022\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://799498681afa7653408bc23b1e82d75c16e3c809d2ee7f322780e20c537eb710\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bdx7g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:23Z\\\"}}\" for pod \"openshift-multus\"/\"multus-kq4z8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:37Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:37 crc kubenswrapper[4980]: I1206 03:34:37.792893 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-6kxfq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"951b6deb-bcea-4332-8ae8-c23ac7aaef36\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:37Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:37Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b8htf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b8htf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:37Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-6kxfq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:37Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:37 crc kubenswrapper[4980]: I1206 03:34:37.805163 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7f958150-549d-4c80-8a7e-23c39e4f9500\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c28b6895b14e73f4f3eb7a8ce93076c05092ba9b58ea0e3e879c65bcfa227a21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://76cff8a2d652df38b04eefc053832baf4cbb5db778ff5e525b18c2825141b118\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://206f6f19fe3197f5d6798cc86c5473bf2decb02a2860ee7ffb5efb5f60952e8c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bafaa9670b05671fb57b658c2960bf3ee32c84e52f86a3a3bbab01d5a38ff0c4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://773a0bc2aab73ffc2717e987543defc9b786cc22300aa54b4ba4f3454504e331\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-06T03:34:22Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1206 03:34:15.305330 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1206 03:34:15.307249 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1592846546/tls.crt::/tmp/serving-cert-1592846546/tls.key\\\\\\\"\\\\nI1206 03:34:21.814037 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1206 03:34:21.816756 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1206 03:34:21.816780 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1206 03:34:21.816806 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1206 03:34:21.816813 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1206 03:34:21.831301 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1206 03:34:21.831311 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1206 03:34:21.831324 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 03:34:21.831330 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 03:34:21.831335 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1206 03:34:21.831339 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1206 03:34:21.831342 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1206 03:34:21.831346 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1206 03:34:21.911269 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:04Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d40e0e7446f8416bf9da94fc53c01be937f9a51a11eb101157d2ddd4597720d2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:04Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2cba89f2f1bed96d59f183628d900d9b72c91eb472e273c8bed0babcab485910\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2cba89f2f1bed96d59f183628d900d9b72c91eb472e273c8bed0babcab485910\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:01Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:37Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:37 crc kubenswrapper[4980]: I1206 03:34:37.817844 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6ed68f66-1cf6-465a-8f66-76d5af6a44d8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://725e84f6bdfb722a8ce69158d03ea4c66a9f087642d55cb9651e003c81be0c8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://71f7b492ea07cdf7dc65fc48184ec1e64914eb1811bbfdddc0fb14c8ad6cb296\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b543aa4ee03d5a14909b67d876c5675c02e1df2a89a3c9ac8d3fb587fea53d8f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://32af3427e872a70cb3c2357116f04cabc8d3d9b69e7306b2e0de2ae22f9469d4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:01Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:37Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:37 crc kubenswrapper[4980]: I1206 03:34:37.828904 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7f958150-549d-4c80-8a7e-23c39e4f9500\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c28b6895b14e73f4f3eb7a8ce93076c05092ba9b58ea0e3e879c65bcfa227a21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://76cff8a2d652df38b04eefc053832baf4cbb5db778ff5e525b18c2825141b118\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://206f6f19fe3197f5d6798cc86c5473bf2decb02a2860ee7ffb5efb5f60952e8c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bafaa9670b05671fb57b658c2960bf3ee32c84e52f86a3a3bbab01d5a38ff0c4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://773a0bc2aab73ffc2717e987543defc9b786cc22300aa54b4ba4f3454504e331\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-06T03:34:22Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1206 03:34:15.305330 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1206 03:34:15.307249 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1592846546/tls.crt::/tmp/serving-cert-1592846546/tls.key\\\\\\\"\\\\nI1206 03:34:21.814037 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1206 03:34:21.816756 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1206 03:34:21.816780 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1206 03:34:21.816806 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1206 03:34:21.816813 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1206 03:34:21.831301 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1206 03:34:21.831311 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1206 03:34:21.831324 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 03:34:21.831330 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 03:34:21.831335 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1206 03:34:21.831339 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1206 03:34:21.831342 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1206 03:34:21.831346 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1206 03:34:21.911269 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:04Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d40e0e7446f8416bf9da94fc53c01be937f9a51a11eb101157d2ddd4597720d2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:04Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2cba89f2f1bed96d59f183628d900d9b72c91eb472e273c8bed0babcab485910\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2cba89f2f1bed96d59f183628d900d9b72c91eb472e273c8bed0babcab485910\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:01Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:37Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:37 crc kubenswrapper[4980]: I1206 03:34:37.835373 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:34:37 crc kubenswrapper[4980]: I1206 03:34:37.835400 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:34:37 crc kubenswrapper[4980]: I1206 03:34:37.835410 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:34:37 crc kubenswrapper[4980]: I1206 03:34:37.835423 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:34:37 crc kubenswrapper[4980]: I1206 03:34:37.835435 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:34:37Z","lastTransitionTime":"2025-12-06T03:34:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:34:37 crc kubenswrapper[4980]: I1206 03:34:37.843819 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6ed68f66-1cf6-465a-8f66-76d5af6a44d8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://725e84f6bdfb722a8ce69158d03ea4c66a9f087642d55cb9651e003c81be0c8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://71f7b492ea07cdf7dc65fc48184ec1e64914eb1811bbfdddc0fb14c8ad6cb296\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b543aa4ee03d5a14909b67d876c5675c02e1df2a89a3c9ac8d3fb587fea53d8f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://32af3427e872a70cb3c2357116f04cabc8d3d9b69e7306b2e0de2ae22f9469d4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:01Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:37Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:37 crc kubenswrapper[4980]: I1206 03:34:37.858015 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:22Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:37Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:37 crc kubenswrapper[4980]: I1206 03:34:37.869267 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://334ea6a273fd7d268906c9f4a6dbf6f4df4fcdbf38f3d7d248cb7d6cf8537794\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ec44dcd5e4bf0a0466bb586ebd61259364dbbfe22499f4cbff854e3bd153082\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:37Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:37 crc kubenswrapper[4980]: I1206 03:34:37.881823 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:22Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:37Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:37 crc kubenswrapper[4980]: I1206 03:34:37.890235 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-sv5sz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"873977aa-4eb0-4526-bc0b-757ad2dea2a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01f0cc855c822a71ea8ef6a7ce6e25747ff9956691735973f935c4a28672a7a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lg7tx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:23Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-sv5sz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:37Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:37 crc kubenswrapper[4980]: I1206 03:34:37.905867 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5w4fr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5af0e768-3c9c-4401-ab56-57516cd3170e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:23Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:23Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://768b8e1fe1a00a8467d744da5ad3f61006b0a3612c96f8e4d0ee6fcb93195f9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfzrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6342903b318317ee8d84c2f5d467cb25c811a789be0d6ab0e0cba37b5611e16d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfzrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d620685574db7594132ae24a91ec0a93dce20a9a289540deebb39650b8a1816\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfzrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c67ec8f9c1764fcdf215d52822c2d4f0437dc7f056f05d708767121f6c1f3515\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfzrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f35beae5339eeaf26494ed678e3c1bd02349eae2a4b4615325b17923f3b653b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfzrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a940ec6ee523313728f7e8ae934b7e0b0cff82771e3e64723309db1fae013c3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfzrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a8d3b4a48fbd8d6d81e7de4c07d3da18b7bfaa209461e10b38daf67000c1f699\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a8d3b4a48fbd8d6d81e7de4c07d3da18b7bfaa209461e10b38daf67000c1f699\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-06T03:34:37Z\\\",\\\"message\\\":\\\":default/a8519615025667110816) with []\\\\nI1206 03:34:37.270920 6168 address_set.go:302] New(bf133528-8652-4c84-85ff-881f0afe9837/default-network-controller:EgressService:egresssvc-served-pods:v4/a13607449821398607916) with []\\\\nI1206 03:34:37.270987 6168 factory.go:1336] Added *v1.Node event handler 7\\\\nI1206 03:34:37.271037 6168 factory.go:1336] Added *v1.EgressIP event handler 8\\\\nI1206 03:34:37.271038 6168 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1206 03:34:37.271088 6168 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1206 03:34:37.271123 6168 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1206 03:34:37.271230 6168 factory.go:656] Stopping watch factory\\\\nI1206 03:34:37.271122 6168 handler.go:208] Removed *v1.Node event handler 2\\\\nI1206 03:34:37.271267 6168 handler.go:208] Removed *v1.Node event handler 7\\\\nI1206 03:34:37.271295 6168 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1206 03:34:37.271581 6168 factory.go:1336] Added *v1.EgressFirewall event handler 9\\\\nI1206 03:34:37.271672 6168 controller.go:132] Adding controller ef_node_controller event handlers\\\\nI1206 03:34:37.271709 6168 ovnkube.go:599] Stopped ovnkube\\\\nI1206 03:34:37.271735 6168 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1206 03:34:37.271795 6168 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfzrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe261f4322b85702181d7473ab2a9a4d2d4b869d586c3091f811b12a81a63aae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfzrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d876df04b5ab0ba07a5cbcafbe4648403c5014c59eb7d40adee2657c3a255473\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d876df04b5ab0ba07a5cbcafbe4648403c5014c59eb7d40adee2657c3a255473\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfzrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:23Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-5w4fr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:37Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:37 crc kubenswrapper[4980]: I1206 03:34:37.924405 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cab608b3-a02f-475b-9aae-9b8f5d878de9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f7fcc44b42bcbb55509048e3a566bb557333c4a83e33d7b0fba8eb514317909\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c22e7dd8886c2dfbb742900ed86ec7b0c2edf9a1363e74f6666961d928f06c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://473e792b2d0a9644854ee0581c6120fc9c989bc342899ef72b1d056080fd4694\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ee8e936dfc39361d6f99460910591ca8c901a20f0d274bbb6aaebfa8a00ee06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://77bc12768b018625ba11a6bfd66688f926c4b7f3bc1d0e41c6551a18b485ecf3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://27cf07cd96893900da4afecd0d56fd9269bc60c1d23246b4685fcd8756357c19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://27cf07cd96893900da4afecd0d56fd9269bc60c1d23246b4685fcd8756357c19\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c49feb24566a8284378670fa0390710d38c237acc94ffb654f90f3fa4626e4c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c49feb24566a8284378670fa0390710d38c237acc94ffb654f90f3fa4626e4c8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:03Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://bb74d1c3d752a379fa394c450217badb45511fbcafad62f7b618131cab94b29e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bb74d1c3d752a379fa394c450217badb45511fbcafad62f7b618131cab94b29e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:01Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:37Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:37 crc kubenswrapper[4980]: I1206 03:34:37.935790 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://502c727d6baa31f400bee14c34b19abbe38af873f28f6a82a43547431755d981\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:37Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:37 crc kubenswrapper[4980]: I1206 03:34:37.938291 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:34:37 crc kubenswrapper[4980]: I1206 03:34:37.938343 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:34:37 crc kubenswrapper[4980]: I1206 03:34:37.938357 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:34:37 crc kubenswrapper[4980]: I1206 03:34:37.938373 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:34:37 crc kubenswrapper[4980]: I1206 03:34:37.938384 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:34:37Z","lastTransitionTime":"2025-12-06T03:34:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:34:37 crc kubenswrapper[4980]: I1206 03:34:37.951203 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-dx44q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5bac22a-512c-49d0-90db-47de6aaaefd0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:23Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:23Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f7mdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3b5dab476c7f227a6e4a2eae8469432667fb9fad345d9419ba59feab9fd385e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c3b5dab476c7f227a6e4a2eae8469432667fb9fad345d9419ba59feab9fd385e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f7mdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e921499231cbdf26b8918c98f9ad8c322fd7848865caeef048470de3e2c0efba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e921499231cbdf26b8918c98f9ad8c322fd7848865caeef048470de3e2c0efba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f7mdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7cbfe7d8cdb89916c238296592fab45b9e92e2c585baa6d67230decd1c84c5bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7cbfe7d8cdb89916c238296592fab45b9e92e2c585baa6d67230decd1c84c5bb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f7mdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f45cbf23b6b0d208cefc4b493f901bab9d71ccf2f1456507d54fb1f30c4fd48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7f45cbf23b6b0d208cefc4b493f901bab9d71ccf2f1456507d54fb1f30c4fd48\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f7mdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://344ce4b145ff021fcaf163803d8ad8cc6e9565bf7f0992e2653c43db663d8411\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://344ce4b145ff021fcaf163803d8ad8cc6e9565bf7f0992e2653c43db663d8411\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f7mdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://619e56a9937dd6376ab471ba16da30e5d77f443f5bd33610bea5ec85cd0446ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://619e56a9937dd6376ab471ba16da30e5d77f443f5bd33610bea5ec85cd0446ef\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f7mdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:23Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-dx44q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:37Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:37 crc kubenswrapper[4980]: I1206 03:34:37.962887 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-r5zfb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"320f44d1-a671-4a91-b328-a8b0fdd8f23a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://92393ecaaef8cfcfd01837671f8cc89803dd570b637b4a373b40b7f487788826\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zhm5t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52e8db644463794041978ced2c0591c13e0ada8e5f1a78a94a0a665218a37b5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zhm5t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:23Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-r5zfb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:37Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:37 crc kubenswrapper[4980]: I1206 03:34:37.973692 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-cvzg8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f1cde706-268d-4d26-818f-2f42bfb37b5f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://86f9a77acae360740be827eb0d1e1516e8a49cb9b197b6f5f9468ac70a101158\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rk86l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:29Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-cvzg8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:37Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:37 crc kubenswrapper[4980]: I1206 03:34:37.983598 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:37Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:37 crc kubenswrapper[4980]: I1206 03:34:37.995659 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4ef01e7d9809fd9cfc46e3a58b07159cdc6f5a3c210d1c9efc49bca27b30858\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:37Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:38 crc kubenswrapper[4980]: I1206 03:34:38.007583 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-kq4z8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f96c1f6c-55e8-4d86-bff3-5381581b7022\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://799498681afa7653408bc23b1e82d75c16e3c809d2ee7f322780e20c537eb710\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bdx7g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:23Z\\\"}}\" for pod \"openshift-multus\"/\"multus-kq4z8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:38Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:38 crc kubenswrapper[4980]: I1206 03:34:38.021864 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-6kxfq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"951b6deb-bcea-4332-8ae8-c23ac7aaef36\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:37Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:37Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b8htf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b8htf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:37Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-6kxfq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:38Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:38 crc kubenswrapper[4980]: I1206 03:34:38.035665 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 03:34:38 crc kubenswrapper[4980]: I1206 03:34:38.035724 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 03:34:38 crc kubenswrapper[4980]: E1206 03:34:38.035814 4980 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 06 03:34:38 crc kubenswrapper[4980]: E1206 03:34:38.035840 4980 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 06 03:34:38 crc kubenswrapper[4980]: E1206 03:34:38.035896 4980 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-06 03:34:54.035876125 +0000 UTC m=+53.280627396 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 06 03:34:38 crc kubenswrapper[4980]: E1206 03:34:38.035930 4980 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-06 03:34:54.035912736 +0000 UTC m=+53.280664007 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 06 03:34:38 crc kubenswrapper[4980]: I1206 03:34:38.041081 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:34:38 crc kubenswrapper[4980]: I1206 03:34:38.041118 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:34:38 crc kubenswrapper[4980]: I1206 03:34:38.041127 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:34:38 crc kubenswrapper[4980]: I1206 03:34:38.041142 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:34:38 crc kubenswrapper[4980]: I1206 03:34:38.041162 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:34:38Z","lastTransitionTime":"2025-12-06T03:34:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:34:38 crc kubenswrapper[4980]: I1206 03:34:38.181524 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 03:34:38 crc kubenswrapper[4980]: E1206 03:34:38.181627 4980 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 03:34:54.181606929 +0000 UTC m=+53.426358200 (durationBeforeRetry 16s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 03:34:38 crc kubenswrapper[4980]: I1206 03:34:38.181669 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 03:34:38 crc kubenswrapper[4980]: I1206 03:34:38.181743 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 03:34:38 crc kubenswrapper[4980]: E1206 03:34:38.181852 4980 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 06 03:34:38 crc kubenswrapper[4980]: E1206 03:34:38.181875 4980 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 06 03:34:38 crc kubenswrapper[4980]: E1206 03:34:38.181876 4980 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 06 03:34:38 crc kubenswrapper[4980]: E1206 03:34:38.181886 4980 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 06 03:34:38 crc kubenswrapper[4980]: E1206 03:34:38.181891 4980 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 06 03:34:38 crc kubenswrapper[4980]: E1206 03:34:38.181901 4980 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 06 03:34:38 crc kubenswrapper[4980]: E1206 03:34:38.181931 4980 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-06 03:34:54.181917067 +0000 UTC m=+53.426668338 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 06 03:34:38 crc kubenswrapper[4980]: E1206 03:34:38.181948 4980 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-06 03:34:54.181942028 +0000 UTC m=+53.426693299 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 06 03:34:38 crc kubenswrapper[4980]: I1206 03:34:38.182973 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:34:38 crc kubenswrapper[4980]: I1206 03:34:38.182999 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:34:38 crc kubenswrapper[4980]: I1206 03:34:38.183008 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:34:38 crc kubenswrapper[4980]: I1206 03:34:38.183022 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:34:38 crc kubenswrapper[4980]: I1206 03:34:38.183032 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:34:38Z","lastTransitionTime":"2025-12-06T03:34:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:34:38 crc kubenswrapper[4980]: I1206 03:34:38.183376 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 03:34:38 crc kubenswrapper[4980]: I1206 03:34:38.183378 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 03:34:38 crc kubenswrapper[4980]: E1206 03:34:38.183454 4980 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 03:34:38 crc kubenswrapper[4980]: E1206 03:34:38.183626 4980 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 03:34:38 crc kubenswrapper[4980]: I1206 03:34:38.285040 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:34:38 crc kubenswrapper[4980]: I1206 03:34:38.285078 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:34:38 crc kubenswrapper[4980]: I1206 03:34:38.285089 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:34:38 crc kubenswrapper[4980]: I1206 03:34:38.285103 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:34:38 crc kubenswrapper[4980]: I1206 03:34:38.285113 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:34:38Z","lastTransitionTime":"2025-12-06T03:34:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:34:38 crc kubenswrapper[4980]: I1206 03:34:38.388451 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:34:38 crc kubenswrapper[4980]: I1206 03:34:38.388570 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:34:38 crc kubenswrapper[4980]: I1206 03:34:38.388599 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:34:38 crc kubenswrapper[4980]: I1206 03:34:38.388633 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:34:38 crc kubenswrapper[4980]: I1206 03:34:38.388653 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:34:38Z","lastTransitionTime":"2025-12-06T03:34:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:34:38 crc kubenswrapper[4980]: I1206 03:34:38.465399 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:34:38 crc kubenswrapper[4980]: I1206 03:34:38.465666 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:34:38 crc kubenswrapper[4980]: I1206 03:34:38.465727 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:34:38 crc kubenswrapper[4980]: I1206 03:34:38.465805 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:34:38 crc kubenswrapper[4980]: I1206 03:34:38.465863 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:34:38Z","lastTransitionTime":"2025-12-06T03:34:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:34:38 crc kubenswrapper[4980]: E1206 03:34:38.482767 4980 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T03:34:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T03:34:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:38Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T03:34:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T03:34:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:38Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"9b4ed4f8-ed8d-447b-9e1f-13defbdcf0f7\\\",\\\"systemUUID\\\":\\\"a5b493aa-69de-4688-8ff6-1e5f49b7f014\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:38Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:38 crc kubenswrapper[4980]: I1206 03:34:38.487332 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:34:38 crc kubenswrapper[4980]: I1206 03:34:38.487402 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:34:38 crc kubenswrapper[4980]: I1206 03:34:38.487413 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:34:38 crc kubenswrapper[4980]: I1206 03:34:38.487436 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:34:38 crc kubenswrapper[4980]: I1206 03:34:38.487448 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:34:38Z","lastTransitionTime":"2025-12-06T03:34:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:34:38 crc kubenswrapper[4980]: E1206 03:34:38.504397 4980 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T03:34:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T03:34:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:38Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T03:34:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T03:34:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:38Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"9b4ed4f8-ed8d-447b-9e1f-13defbdcf0f7\\\",\\\"systemUUID\\\":\\\"a5b493aa-69de-4688-8ff6-1e5f49b7f014\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:38Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:38 crc kubenswrapper[4980]: I1206 03:34:38.508927 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:34:38 crc kubenswrapper[4980]: I1206 03:34:38.508989 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:34:38 crc kubenswrapper[4980]: I1206 03:34:38.509009 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:34:38 crc kubenswrapper[4980]: I1206 03:34:38.509032 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:34:38 crc kubenswrapper[4980]: I1206 03:34:38.509048 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:34:38Z","lastTransitionTime":"2025-12-06T03:34:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:34:38 crc kubenswrapper[4980]: E1206 03:34:38.523198 4980 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T03:34:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T03:34:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:38Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T03:34:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T03:34:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:38Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"9b4ed4f8-ed8d-447b-9e1f-13defbdcf0f7\\\",\\\"systemUUID\\\":\\\"a5b493aa-69de-4688-8ff6-1e5f49b7f014\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:38Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:38 crc kubenswrapper[4980]: I1206 03:34:38.527061 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:34:38 crc kubenswrapper[4980]: I1206 03:34:38.527174 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:34:38 crc kubenswrapper[4980]: I1206 03:34:38.527234 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:34:38 crc kubenswrapper[4980]: I1206 03:34:38.527305 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:34:38 crc kubenswrapper[4980]: I1206 03:34:38.527361 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:34:38Z","lastTransitionTime":"2025-12-06T03:34:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:34:38 crc kubenswrapper[4980]: E1206 03:34:38.544614 4980 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T03:34:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T03:34:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:38Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T03:34:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T03:34:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:38Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"9b4ed4f8-ed8d-447b-9e1f-13defbdcf0f7\\\",\\\"systemUUID\\\":\\\"a5b493aa-69de-4688-8ff6-1e5f49b7f014\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:38Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:38 crc kubenswrapper[4980]: I1206 03:34:38.548961 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:34:38 crc kubenswrapper[4980]: I1206 03:34:38.548996 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:34:38 crc kubenswrapper[4980]: I1206 03:34:38.549009 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:34:38 crc kubenswrapper[4980]: I1206 03:34:38.549027 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:34:38 crc kubenswrapper[4980]: I1206 03:34:38.549039 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:34:38Z","lastTransitionTime":"2025-12-06T03:34:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:34:38 crc kubenswrapper[4980]: I1206 03:34:38.555180 4980 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/network-metrics-daemon-mcprh"] Dec 06 03:34:38 crc kubenswrapper[4980]: I1206 03:34:38.555703 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-mcprh" Dec 06 03:34:38 crc kubenswrapper[4980]: E1206 03:34:38.555775 4980 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-mcprh" podUID="e6e88c5e-a255-449a-ae38-ac4d73a8e984" Dec 06 03:34:38 crc kubenswrapper[4980]: I1206 03:34:38.577700 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7f958150-549d-4c80-8a7e-23c39e4f9500\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c28b6895b14e73f4f3eb7a8ce93076c05092ba9b58ea0e3e879c65bcfa227a21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://76cff8a2d652df38b04eefc053832baf4cbb5db778ff5e525b18c2825141b118\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://206f6f19fe3197f5d6798cc86c5473bf2decb02a2860ee7ffb5efb5f60952e8c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bafaa9670b05671fb57b658c2960bf3ee32c84e52f86a3a3bbab01d5a38ff0c4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://773a0bc2aab73ffc2717e987543defc9b786cc22300aa54b4ba4f3454504e331\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-06T03:34:22Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1206 03:34:15.305330 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1206 03:34:15.307249 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1592846546/tls.crt::/tmp/serving-cert-1592846546/tls.key\\\\\\\"\\\\nI1206 03:34:21.814037 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1206 03:34:21.816756 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1206 03:34:21.816780 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1206 03:34:21.816806 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1206 03:34:21.816813 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1206 03:34:21.831301 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1206 03:34:21.831311 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1206 03:34:21.831324 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 03:34:21.831330 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 03:34:21.831335 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1206 03:34:21.831339 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1206 03:34:21.831342 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1206 03:34:21.831346 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1206 03:34:21.911269 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:04Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d40e0e7446f8416bf9da94fc53c01be937f9a51a11eb101157d2ddd4597720d2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:04Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2cba89f2f1bed96d59f183628d900d9b72c91eb472e273c8bed0babcab485910\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2cba89f2f1bed96d59f183628d900d9b72c91eb472e273c8bed0babcab485910\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:01Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:38Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:38 crc kubenswrapper[4980]: E1206 03:34:38.580936 4980 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T03:34:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T03:34:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:38Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T03:34:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T03:34:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:38Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"9b4ed4f8-ed8d-447b-9e1f-13defbdcf0f7\\\",\\\"systemUUID\\\":\\\"a5b493aa-69de-4688-8ff6-1e5f49b7f014\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:38Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:38 crc kubenswrapper[4980]: E1206 03:34:38.581255 4980 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 06 03:34:38 crc kubenswrapper[4980]: I1206 03:34:38.583923 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:34:38 crc kubenswrapper[4980]: I1206 03:34:38.583996 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:34:38 crc kubenswrapper[4980]: I1206 03:34:38.584010 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:34:38 crc kubenswrapper[4980]: I1206 03:34:38.584102 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:34:38 crc kubenswrapper[4980]: I1206 03:34:38.584126 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:34:38Z","lastTransitionTime":"2025-12-06T03:34:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:34:38 crc kubenswrapper[4980]: I1206 03:34:38.591398 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6ed68f66-1cf6-465a-8f66-76d5af6a44d8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://725e84f6bdfb722a8ce69158d03ea4c66a9f087642d55cb9651e003c81be0c8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://71f7b492ea07cdf7dc65fc48184ec1e64914eb1811bbfdddc0fb14c8ad6cb296\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b543aa4ee03d5a14909b67d876c5675c02e1df2a89a3c9ac8d3fb587fea53d8f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://32af3427e872a70cb3c2357116f04cabc8d3d9b69e7306b2e0de2ae22f9469d4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:01Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:38Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:38 crc kubenswrapper[4980]: I1206 03:34:38.603116 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:22Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:38Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:38 crc kubenswrapper[4980]: I1206 03:34:38.614108 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-sv5sz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"873977aa-4eb0-4526-bc0b-757ad2dea2a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01f0cc855c822a71ea8ef6a7ce6e25747ff9956691735973f935c4a28672a7a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lg7tx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:23Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-sv5sz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:38Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:38 crc kubenswrapper[4980]: I1206 03:34:38.637183 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5w4fr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5af0e768-3c9c-4401-ab56-57516cd3170e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:23Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:23Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://768b8e1fe1a00a8467d744da5ad3f61006b0a3612c96f8e4d0ee6fcb93195f9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfzrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6342903b318317ee8d84c2f5d467cb25c811a789be0d6ab0e0cba37b5611e16d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfzrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d620685574db7594132ae24a91ec0a93dce20a9a289540deebb39650b8a1816\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfzrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c67ec8f9c1764fcdf215d52822c2d4f0437dc7f056f05d708767121f6c1f3515\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfzrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f35beae5339eeaf26494ed678e3c1bd02349eae2a4b4615325b17923f3b653b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfzrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a940ec6ee523313728f7e8ae934b7e0b0cff82771e3e64723309db1fae013c3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfzrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a8d3b4a48fbd8d6d81e7de4c07d3da18b7bfaa209461e10b38daf67000c1f699\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a8d3b4a48fbd8d6d81e7de4c07d3da18b7bfaa209461e10b38daf67000c1f699\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-06T03:34:37Z\\\",\\\"message\\\":\\\":default/a8519615025667110816) with []\\\\nI1206 03:34:37.270920 6168 address_set.go:302] New(bf133528-8652-4c84-85ff-881f0afe9837/default-network-controller:EgressService:egresssvc-served-pods:v4/a13607449821398607916) with []\\\\nI1206 03:34:37.270987 6168 factory.go:1336] Added *v1.Node event handler 7\\\\nI1206 03:34:37.271037 6168 factory.go:1336] Added *v1.EgressIP event handler 8\\\\nI1206 03:34:37.271038 6168 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1206 03:34:37.271088 6168 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1206 03:34:37.271123 6168 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1206 03:34:37.271230 6168 factory.go:656] Stopping watch factory\\\\nI1206 03:34:37.271122 6168 handler.go:208] Removed *v1.Node event handler 2\\\\nI1206 03:34:37.271267 6168 handler.go:208] Removed *v1.Node event handler 7\\\\nI1206 03:34:37.271295 6168 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1206 03:34:37.271581 6168 factory.go:1336] Added *v1.EgressFirewall event handler 9\\\\nI1206 03:34:37.271672 6168 controller.go:132] Adding controller ef_node_controller event handlers\\\\nI1206 03:34:37.271709 6168 ovnkube.go:599] Stopped ovnkube\\\\nI1206 03:34:37.271735 6168 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1206 03:34:37.271795 6168 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfzrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe261f4322b85702181d7473ab2a9a4d2d4b869d586c3091f811b12a81a63aae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfzrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d876df04b5ab0ba07a5cbcafbe4648403c5014c59eb7d40adee2657c3a255473\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d876df04b5ab0ba07a5cbcafbe4648403c5014c59eb7d40adee2657c3a255473\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfzrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:23Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-5w4fr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:38Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:38 crc kubenswrapper[4980]: I1206 03:34:38.656193 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:22Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:38Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:38 crc kubenswrapper[4980]: I1206 03:34:38.676573 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://334ea6a273fd7d268906c9f4a6dbf6f4df4fcdbf38f3d7d248cb7d6cf8537794\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ec44dcd5e4bf0a0466bb586ebd61259364dbbfe22499f4cbff854e3bd153082\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:38Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:38 crc kubenswrapper[4980]: I1206 03:34:38.686860 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/e6e88c5e-a255-449a-ae38-ac4d73a8e984-metrics-certs\") pod \"network-metrics-daemon-mcprh\" (UID: \"e6e88c5e-a255-449a-ae38-ac4d73a8e984\") " pod="openshift-multus/network-metrics-daemon-mcprh" Dec 06 03:34:38 crc kubenswrapper[4980]: I1206 03:34:38.687080 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g7xz9\" (UniqueName: \"kubernetes.io/projected/e6e88c5e-a255-449a-ae38-ac4d73a8e984-kube-api-access-g7xz9\") pod \"network-metrics-daemon-mcprh\" (UID: \"e6e88c5e-a255-449a-ae38-ac4d73a8e984\") " pod="openshift-multus/network-metrics-daemon-mcprh" Dec 06 03:34:38 crc kubenswrapper[4980]: I1206 03:34:38.687153 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:34:38 crc kubenswrapper[4980]: I1206 03:34:38.687187 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:34:38 crc kubenswrapper[4980]: I1206 03:34:38.687197 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:34:38 crc kubenswrapper[4980]: I1206 03:34:38.687220 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:34:38 crc kubenswrapper[4980]: I1206 03:34:38.687236 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:34:38Z","lastTransitionTime":"2025-12-06T03:34:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:34:38 crc kubenswrapper[4980]: I1206 03:34:38.694215 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-r5zfb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"320f44d1-a671-4a91-b328-a8b0fdd8f23a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://92393ecaaef8cfcfd01837671f8cc89803dd570b637b4a373b40b7f487788826\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zhm5t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52e8db644463794041978ced2c0591c13e0ada8e5f1a78a94a0a665218a37b5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zhm5t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:23Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-r5zfb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:38Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:38 crc kubenswrapper[4980]: I1206 03:34:38.706100 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-cvzg8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f1cde706-268d-4d26-818f-2f42bfb37b5f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://86f9a77acae360740be827eb0d1e1516e8a49cb9b197b6f5f9468ac70a101158\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rk86l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:29Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-cvzg8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:38Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:38 crc kubenswrapper[4980]: I1206 03:34:38.727025 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cab608b3-a02f-475b-9aae-9b8f5d878de9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f7fcc44b42bcbb55509048e3a566bb557333c4a83e33d7b0fba8eb514317909\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c22e7dd8886c2dfbb742900ed86ec7b0c2edf9a1363e74f6666961d928f06c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://473e792b2d0a9644854ee0581c6120fc9c989bc342899ef72b1d056080fd4694\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ee8e936dfc39361d6f99460910591ca8c901a20f0d274bbb6aaebfa8a00ee06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://77bc12768b018625ba11a6bfd66688f926c4b7f3bc1d0e41c6551a18b485ecf3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://27cf07cd96893900da4afecd0d56fd9269bc60c1d23246b4685fcd8756357c19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://27cf07cd96893900da4afecd0d56fd9269bc60c1d23246b4685fcd8756357c19\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c49feb24566a8284378670fa0390710d38c237acc94ffb654f90f3fa4626e4c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c49feb24566a8284378670fa0390710d38c237acc94ffb654f90f3fa4626e4c8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:03Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://bb74d1c3d752a379fa394c450217badb45511fbcafad62f7b618131cab94b29e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bb74d1c3d752a379fa394c450217badb45511fbcafad62f7b618131cab94b29e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:01Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:38Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:38 crc kubenswrapper[4980]: I1206 03:34:38.741209 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://502c727d6baa31f400bee14c34b19abbe38af873f28f6a82a43547431755d981\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:38Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:38 crc kubenswrapper[4980]: I1206 03:34:38.760219 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-dx44q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5bac22a-512c-49d0-90db-47de6aaaefd0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:23Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:23Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f7mdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3b5dab476c7f227a6e4a2eae8469432667fb9fad345d9419ba59feab9fd385e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c3b5dab476c7f227a6e4a2eae8469432667fb9fad345d9419ba59feab9fd385e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f7mdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e921499231cbdf26b8918c98f9ad8c322fd7848865caeef048470de3e2c0efba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e921499231cbdf26b8918c98f9ad8c322fd7848865caeef048470de3e2c0efba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f7mdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7cbfe7d8cdb89916c238296592fab45b9e92e2c585baa6d67230decd1c84c5bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7cbfe7d8cdb89916c238296592fab45b9e92e2c585baa6d67230decd1c84c5bb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f7mdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f45cbf23b6b0d208cefc4b493f901bab9d71ccf2f1456507d54fb1f30c4fd48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7f45cbf23b6b0d208cefc4b493f901bab9d71ccf2f1456507d54fb1f30c4fd48\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f7mdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://344ce4b145ff021fcaf163803d8ad8cc6e9565bf7f0992e2653c43db663d8411\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://344ce4b145ff021fcaf163803d8ad8cc6e9565bf7f0992e2653c43db663d8411\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f7mdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://619e56a9937dd6376ab471ba16da30e5d77f443f5bd33610bea5ec85cd0446ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://619e56a9937dd6376ab471ba16da30e5d77f443f5bd33610bea5ec85cd0446ef\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f7mdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:23Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-dx44q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:38Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:38 crc kubenswrapper[4980]: I1206 03:34:38.774853 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-6kxfq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"951b6deb-bcea-4332-8ae8-c23ac7aaef36\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:37Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:37Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b8htf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b8htf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:37Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-6kxfq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:38Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:38 crc kubenswrapper[4980]: I1206 03:34:38.789027 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g7xz9\" (UniqueName: \"kubernetes.io/projected/e6e88c5e-a255-449a-ae38-ac4d73a8e984-kube-api-access-g7xz9\") pod \"network-metrics-daemon-mcprh\" (UID: \"e6e88c5e-a255-449a-ae38-ac4d73a8e984\") " pod="openshift-multus/network-metrics-daemon-mcprh" Dec 06 03:34:38 crc kubenswrapper[4980]: I1206 03:34:38.789089 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/e6e88c5e-a255-449a-ae38-ac4d73a8e984-metrics-certs\") pod \"network-metrics-daemon-mcprh\" (UID: \"e6e88c5e-a255-449a-ae38-ac4d73a8e984\") " pod="openshift-multus/network-metrics-daemon-mcprh" Dec 06 03:34:38 crc kubenswrapper[4980]: E1206 03:34:38.789214 4980 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 06 03:34:38 crc kubenswrapper[4980]: E1206 03:34:38.789280 4980 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/e6e88c5e-a255-449a-ae38-ac4d73a8e984-metrics-certs podName:e6e88c5e-a255-449a-ae38-ac4d73a8e984 nodeName:}" failed. No retries permitted until 2025-12-06 03:34:39.289263183 +0000 UTC m=+38.534014454 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/e6e88c5e-a255-449a-ae38-ac4d73a8e984-metrics-certs") pod "network-metrics-daemon-mcprh" (UID: "e6e88c5e-a255-449a-ae38-ac4d73a8e984") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 06 03:34:38 crc kubenswrapper[4980]: I1206 03:34:38.791598 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-mcprh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e6e88c5e-a255-449a-ae38-ac4d73a8e984\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g7xz9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g7xz9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:38Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-mcprh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:38Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:38 crc kubenswrapper[4980]: I1206 03:34:38.791783 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:34:38 crc kubenswrapper[4980]: I1206 03:34:38.791889 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:34:38 crc kubenswrapper[4980]: I1206 03:34:38.791948 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:34:38 crc kubenswrapper[4980]: I1206 03:34:38.791976 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:34:38 crc kubenswrapper[4980]: I1206 03:34:38.792032 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:34:38Z","lastTransitionTime":"2025-12-06T03:34:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:34:38 crc kubenswrapper[4980]: I1206 03:34:38.807893 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:38Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:38 crc kubenswrapper[4980]: I1206 03:34:38.812129 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g7xz9\" (UniqueName: \"kubernetes.io/projected/e6e88c5e-a255-449a-ae38-ac4d73a8e984-kube-api-access-g7xz9\") pod \"network-metrics-daemon-mcprh\" (UID: \"e6e88c5e-a255-449a-ae38-ac4d73a8e984\") " pod="openshift-multus/network-metrics-daemon-mcprh" Dec 06 03:34:38 crc kubenswrapper[4980]: I1206 03:34:38.826224 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4ef01e7d9809fd9cfc46e3a58b07159cdc6f5a3c210d1c9efc49bca27b30858\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:38Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:38 crc kubenswrapper[4980]: I1206 03:34:38.841308 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-kq4z8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f96c1f6c-55e8-4d86-bff3-5381581b7022\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://799498681afa7653408bc23b1e82d75c16e3c809d2ee7f322780e20c537eb710\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bdx7g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:23Z\\\"}}\" for pod \"openshift-multus\"/\"multus-kq4z8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:38Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:38 crc kubenswrapper[4980]: I1206 03:34:38.879272 4980 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-5w4fr_5af0e768-3c9c-4401-ab56-57516cd3170e/ovnkube-controller/0.log" Dec 06 03:34:38 crc kubenswrapper[4980]: I1206 03:34:38.882847 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5w4fr" event={"ID":"5af0e768-3c9c-4401-ab56-57516cd3170e","Type":"ContainerStarted","Data":"6cb5cc9f6f22133cebf095db595e5062d799d7e3c8ca0bf31d3809a296e507cc"} Dec 06 03:34:38 crc kubenswrapper[4980]: I1206 03:34:38.883398 4980 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-5w4fr" Dec 06 03:34:38 crc kubenswrapper[4980]: I1206 03:34:38.886443 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-6kxfq" event={"ID":"951b6deb-bcea-4332-8ae8-c23ac7aaef36","Type":"ContainerStarted","Data":"423d30c3e65b04bdf6ff5ef53d8a7575a434c05e28489370a7370381ebc29de9"} Dec 06 03:34:38 crc kubenswrapper[4980]: I1206 03:34:38.890321 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-dx44q" event={"ID":"e5bac22a-512c-49d0-90db-47de6aaaefd0","Type":"ContainerStarted","Data":"97469dea2cd5393da0acf784ccc1969c8a4d07da7d51b14cdc955013ee9710ad"} Dec 06 03:34:38 crc kubenswrapper[4980]: I1206 03:34:38.895809 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:34:38 crc kubenswrapper[4980]: I1206 03:34:38.895850 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:34:38 crc kubenswrapper[4980]: I1206 03:34:38.895865 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:34:38 crc kubenswrapper[4980]: I1206 03:34:38.895885 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:34:38 crc kubenswrapper[4980]: I1206 03:34:38.895902 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:34:38Z","lastTransitionTime":"2025-12-06T03:34:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:34:38 crc kubenswrapper[4980]: I1206 03:34:38.900234 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-dx44q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5bac22a-512c-49d0-90db-47de6aaaefd0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:23Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:23Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f7mdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3b5dab476c7f227a6e4a2eae8469432667fb9fad345d9419ba59feab9fd385e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c3b5dab476c7f227a6e4a2eae8469432667fb9fad345d9419ba59feab9fd385e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f7mdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e921499231cbdf26b8918c98f9ad8c322fd7848865caeef048470de3e2c0efba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e921499231cbdf26b8918c98f9ad8c322fd7848865caeef048470de3e2c0efba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f7mdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7cbfe7d8cdb89916c238296592fab45b9e92e2c585baa6d67230decd1c84c5bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7cbfe7d8cdb89916c238296592fab45b9e92e2c585baa6d67230decd1c84c5bb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f7mdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f45cbf23b6b0d208cefc4b493f901bab9d71ccf2f1456507d54fb1f30c4fd48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7f45cbf23b6b0d208cefc4b493f901bab9d71ccf2f1456507d54fb1f30c4fd48\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f7mdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://344ce4b145ff021fcaf163803d8ad8cc6e9565bf7f0992e2653c43db663d8411\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://344ce4b145ff021fcaf163803d8ad8cc6e9565bf7f0992e2653c43db663d8411\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f7mdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://619e56a9937dd6376ab471ba16da30e5d77f443f5bd33610bea5ec85cd0446ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://619e56a9937dd6376ab471ba16da30e5d77f443f5bd33610bea5ec85cd0446ef\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f7mdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:23Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-dx44q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:38Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:38 crc kubenswrapper[4980]: I1206 03:34:38.926448 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-r5zfb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"320f44d1-a671-4a91-b328-a8b0fdd8f23a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://92393ecaaef8cfcfd01837671f8cc89803dd570b637b4a373b40b7f487788826\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zhm5t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52e8db644463794041978ced2c0591c13e0ada8e5f1a78a94a0a665218a37b5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zhm5t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:23Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-r5zfb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:38Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:38 crc kubenswrapper[4980]: I1206 03:34:38.936657 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-cvzg8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f1cde706-268d-4d26-818f-2f42bfb37b5f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://86f9a77acae360740be827eb0d1e1516e8a49cb9b197b6f5f9468ac70a101158\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rk86l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:29Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-cvzg8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:38Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:38 crc kubenswrapper[4980]: I1206 03:34:38.956198 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cab608b3-a02f-475b-9aae-9b8f5d878de9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f7fcc44b42bcbb55509048e3a566bb557333c4a83e33d7b0fba8eb514317909\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c22e7dd8886c2dfbb742900ed86ec7b0c2edf9a1363e74f6666961d928f06c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://473e792b2d0a9644854ee0581c6120fc9c989bc342899ef72b1d056080fd4694\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ee8e936dfc39361d6f99460910591ca8c901a20f0d274bbb6aaebfa8a00ee06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://77bc12768b018625ba11a6bfd66688f926c4b7f3bc1d0e41c6551a18b485ecf3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://27cf07cd96893900da4afecd0d56fd9269bc60c1d23246b4685fcd8756357c19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://27cf07cd96893900da4afecd0d56fd9269bc60c1d23246b4685fcd8756357c19\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c49feb24566a8284378670fa0390710d38c237acc94ffb654f90f3fa4626e4c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c49feb24566a8284378670fa0390710d38c237acc94ffb654f90f3fa4626e4c8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:03Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://bb74d1c3d752a379fa394c450217badb45511fbcafad62f7b618131cab94b29e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bb74d1c3d752a379fa394c450217badb45511fbcafad62f7b618131cab94b29e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:01Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:38Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:38 crc kubenswrapper[4980]: I1206 03:34:38.967299 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://502c727d6baa31f400bee14c34b19abbe38af873f28f6a82a43547431755d981\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:38Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:38 crc kubenswrapper[4980]: I1206 03:34:38.978637 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-kq4z8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f96c1f6c-55e8-4d86-bff3-5381581b7022\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://799498681afa7653408bc23b1e82d75c16e3c809d2ee7f322780e20c537eb710\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bdx7g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:23Z\\\"}}\" for pod \"openshift-multus\"/\"multus-kq4z8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:38Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:38 crc kubenswrapper[4980]: I1206 03:34:38.988857 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-6kxfq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"951b6deb-bcea-4332-8ae8-c23ac7aaef36\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:37Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:37Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b8htf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b8htf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:37Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-6kxfq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:38Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:38 crc kubenswrapper[4980]: I1206 03:34:38.998211 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:34:38 crc kubenswrapper[4980]: I1206 03:34:38.998250 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:34:38 crc kubenswrapper[4980]: I1206 03:34:38.998261 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:34:38 crc kubenswrapper[4980]: I1206 03:34:38.998280 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:34:38 crc kubenswrapper[4980]: I1206 03:34:38.998291 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:34:38Z","lastTransitionTime":"2025-12-06T03:34:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:34:39 crc kubenswrapper[4980]: I1206 03:34:39.009593 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-mcprh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e6e88c5e-a255-449a-ae38-ac4d73a8e984\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g7xz9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g7xz9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:38Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-mcprh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:39Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:39 crc kubenswrapper[4980]: I1206 03:34:39.023750 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:39Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:39 crc kubenswrapper[4980]: I1206 03:34:39.038560 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4ef01e7d9809fd9cfc46e3a58b07159cdc6f5a3c210d1c9efc49bca27b30858\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:39Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:39 crc kubenswrapper[4980]: I1206 03:34:39.050085 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6ed68f66-1cf6-465a-8f66-76d5af6a44d8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://725e84f6bdfb722a8ce69158d03ea4c66a9f087642d55cb9651e003c81be0c8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://71f7b492ea07cdf7dc65fc48184ec1e64914eb1811bbfdddc0fb14c8ad6cb296\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b543aa4ee03d5a14909b67d876c5675c02e1df2a89a3c9ac8d3fb587fea53d8f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://32af3427e872a70cb3c2357116f04cabc8d3d9b69e7306b2e0de2ae22f9469d4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:01Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:39Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:39 crc kubenswrapper[4980]: I1206 03:34:39.063135 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7f958150-549d-4c80-8a7e-23c39e4f9500\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c28b6895b14e73f4f3eb7a8ce93076c05092ba9b58ea0e3e879c65bcfa227a21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://76cff8a2d652df38b04eefc053832baf4cbb5db778ff5e525b18c2825141b118\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://206f6f19fe3197f5d6798cc86c5473bf2decb02a2860ee7ffb5efb5f60952e8c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bafaa9670b05671fb57b658c2960bf3ee32c84e52f86a3a3bbab01d5a38ff0c4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://773a0bc2aab73ffc2717e987543defc9b786cc22300aa54b4ba4f3454504e331\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-06T03:34:22Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1206 03:34:15.305330 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1206 03:34:15.307249 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1592846546/tls.crt::/tmp/serving-cert-1592846546/tls.key\\\\\\\"\\\\nI1206 03:34:21.814037 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1206 03:34:21.816756 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1206 03:34:21.816780 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1206 03:34:21.816806 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1206 03:34:21.816813 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1206 03:34:21.831301 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1206 03:34:21.831311 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1206 03:34:21.831324 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 03:34:21.831330 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 03:34:21.831335 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1206 03:34:21.831339 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1206 03:34:21.831342 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1206 03:34:21.831346 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1206 03:34:21.911269 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:04Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d40e0e7446f8416bf9da94fc53c01be937f9a51a11eb101157d2ddd4597720d2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:04Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2cba89f2f1bed96d59f183628d900d9b72c91eb472e273c8bed0babcab485910\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2cba89f2f1bed96d59f183628d900d9b72c91eb472e273c8bed0babcab485910\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:01Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:39Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:39 crc kubenswrapper[4980]: I1206 03:34:39.075266 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://334ea6a273fd7d268906c9f4a6dbf6f4df4fcdbf38f3d7d248cb7d6cf8537794\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ec44dcd5e4bf0a0466bb586ebd61259364dbbfe22499f4cbff854e3bd153082\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:39Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:39 crc kubenswrapper[4980]: I1206 03:34:39.091047 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:22Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:39Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:39 crc kubenswrapper[4980]: I1206 03:34:39.100617 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:34:39 crc kubenswrapper[4980]: I1206 03:34:39.100647 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:34:39 crc kubenswrapper[4980]: I1206 03:34:39.100655 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:34:39 crc kubenswrapper[4980]: I1206 03:34:39.100670 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:34:39 crc kubenswrapper[4980]: I1206 03:34:39.100682 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:34:39Z","lastTransitionTime":"2025-12-06T03:34:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:34:39 crc kubenswrapper[4980]: I1206 03:34:39.102940 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-sv5sz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"873977aa-4eb0-4526-bc0b-757ad2dea2a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01f0cc855c822a71ea8ef6a7ce6e25747ff9956691735973f935c4a28672a7a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lg7tx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:23Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-sv5sz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:39Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:39 crc kubenswrapper[4980]: I1206 03:34:39.121126 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5w4fr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5af0e768-3c9c-4401-ab56-57516cd3170e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:23Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:23Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://768b8e1fe1a00a8467d744da5ad3f61006b0a3612c96f8e4d0ee6fcb93195f9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfzrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6342903b318317ee8d84c2f5d467cb25c811a789be0d6ab0e0cba37b5611e16d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfzrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d620685574db7594132ae24a91ec0a93dce20a9a289540deebb39650b8a1816\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfzrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c67ec8f9c1764fcdf215d52822c2d4f0437dc7f056f05d708767121f6c1f3515\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfzrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f35beae5339eeaf26494ed678e3c1bd02349eae2a4b4615325b17923f3b653b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfzrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a940ec6ee523313728f7e8ae934b7e0b0cff82771e3e64723309db1fae013c3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfzrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6cb5cc9f6f22133cebf095db595e5062d799d7e3c8ca0bf31d3809a296e507cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a8d3b4a48fbd8d6d81e7de4c07d3da18b7bfaa209461e10b38daf67000c1f699\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-06T03:34:37Z\\\",\\\"message\\\":\\\":default/a8519615025667110816) with []\\\\nI1206 03:34:37.270920 6168 address_set.go:302] New(bf133528-8652-4c84-85ff-881f0afe9837/default-network-controller:EgressService:egresssvc-served-pods:v4/a13607449821398607916) with []\\\\nI1206 03:34:37.270987 6168 factory.go:1336] Added *v1.Node event handler 7\\\\nI1206 03:34:37.271037 6168 factory.go:1336] Added *v1.EgressIP event handler 8\\\\nI1206 03:34:37.271038 6168 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1206 03:34:37.271088 6168 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1206 03:34:37.271123 6168 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1206 03:34:37.271230 6168 factory.go:656] Stopping watch factory\\\\nI1206 03:34:37.271122 6168 handler.go:208] Removed *v1.Node event handler 2\\\\nI1206 03:34:37.271267 6168 handler.go:208] Removed *v1.Node event handler 7\\\\nI1206 03:34:37.271295 6168 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1206 03:34:37.271581 6168 factory.go:1336] Added *v1.EgressFirewall event handler 9\\\\nI1206 03:34:37.271672 6168 controller.go:132] Adding controller ef_node_controller event handlers\\\\nI1206 03:34:37.271709 6168 ovnkube.go:599] Stopped ovnkube\\\\nI1206 03:34:37.271735 6168 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1206 03:34:37.271795 6168 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:32Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfzrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe261f4322b85702181d7473ab2a9a4d2d4b869d586c3091f811b12a81a63aae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfzrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d876df04b5ab0ba07a5cbcafbe4648403c5014c59eb7d40adee2657c3a255473\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d876df04b5ab0ba07a5cbcafbe4648403c5014c59eb7d40adee2657c3a255473\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfzrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:23Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-5w4fr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:39Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:39 crc kubenswrapper[4980]: I1206 03:34:39.139615 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:22Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:39Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:39 crc kubenswrapper[4980]: I1206 03:34:39.152976 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-sv5sz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"873977aa-4eb0-4526-bc0b-757ad2dea2a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01f0cc855c822a71ea8ef6a7ce6e25747ff9956691735973f935c4a28672a7a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lg7tx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:23Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-sv5sz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:39Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:39 crc kubenswrapper[4980]: I1206 03:34:39.171997 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5w4fr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5af0e768-3c9c-4401-ab56-57516cd3170e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:23Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:23Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://768b8e1fe1a00a8467d744da5ad3f61006b0a3612c96f8e4d0ee6fcb93195f9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfzrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6342903b318317ee8d84c2f5d467cb25c811a789be0d6ab0e0cba37b5611e16d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfzrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d620685574db7594132ae24a91ec0a93dce20a9a289540deebb39650b8a1816\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfzrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c67ec8f9c1764fcdf215d52822c2d4f0437dc7f056f05d708767121f6c1f3515\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfzrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f35beae5339eeaf26494ed678e3c1bd02349eae2a4b4615325b17923f3b653b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfzrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a940ec6ee523313728f7e8ae934b7e0b0cff82771e3e64723309db1fae013c3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfzrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6cb5cc9f6f22133cebf095db595e5062d799d7e3c8ca0bf31d3809a296e507cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a8d3b4a48fbd8d6d81e7de4c07d3da18b7bfaa209461e10b38daf67000c1f699\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-06T03:34:37Z\\\",\\\"message\\\":\\\":default/a8519615025667110816) with []\\\\nI1206 03:34:37.270920 6168 address_set.go:302] New(bf133528-8652-4c84-85ff-881f0afe9837/default-network-controller:EgressService:egresssvc-served-pods:v4/a13607449821398607916) with []\\\\nI1206 03:34:37.270987 6168 factory.go:1336] Added *v1.Node event handler 7\\\\nI1206 03:34:37.271037 6168 factory.go:1336] Added *v1.EgressIP event handler 8\\\\nI1206 03:34:37.271038 6168 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1206 03:34:37.271088 6168 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1206 03:34:37.271123 6168 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1206 03:34:37.271230 6168 factory.go:656] Stopping watch factory\\\\nI1206 03:34:37.271122 6168 handler.go:208] Removed *v1.Node event handler 2\\\\nI1206 03:34:37.271267 6168 handler.go:208] Removed *v1.Node event handler 7\\\\nI1206 03:34:37.271295 6168 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1206 03:34:37.271581 6168 factory.go:1336] Added *v1.EgressFirewall event handler 9\\\\nI1206 03:34:37.271672 6168 controller.go:132] Adding controller ef_node_controller event handlers\\\\nI1206 03:34:37.271709 6168 ovnkube.go:599] Stopped ovnkube\\\\nI1206 03:34:37.271735 6168 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1206 03:34:37.271795 6168 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:32Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfzrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe261f4322b85702181d7473ab2a9a4d2d4b869d586c3091f811b12a81a63aae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfzrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d876df04b5ab0ba07a5cbcafbe4648403c5014c59eb7d40adee2657c3a255473\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d876df04b5ab0ba07a5cbcafbe4648403c5014c59eb7d40adee2657c3a255473\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfzrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:23Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-5w4fr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:39Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:39 crc kubenswrapper[4980]: I1206 03:34:39.183982 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 03:34:39 crc kubenswrapper[4980]: E1206 03:34:39.184131 4980 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 03:34:39 crc kubenswrapper[4980]: I1206 03:34:39.187292 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:22Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:39Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:39 crc kubenswrapper[4980]: I1206 03:34:39.203117 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:34:39 crc kubenswrapper[4980]: I1206 03:34:39.203146 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:34:39 crc kubenswrapper[4980]: I1206 03:34:39.203154 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:34:39 crc kubenswrapper[4980]: I1206 03:34:39.203167 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:34:39 crc kubenswrapper[4980]: I1206 03:34:39.203175 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:34:39Z","lastTransitionTime":"2025-12-06T03:34:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:34:39 crc kubenswrapper[4980]: I1206 03:34:39.203760 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://334ea6a273fd7d268906c9f4a6dbf6f4df4fcdbf38f3d7d248cb7d6cf8537794\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ec44dcd5e4bf0a0466bb586ebd61259364dbbfe22499f4cbff854e3bd153082\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:39Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:39 crc kubenswrapper[4980]: I1206 03:34:39.249459 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:22Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:39Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:39 crc kubenswrapper[4980]: I1206 03:34:39.263260 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-cvzg8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f1cde706-268d-4d26-818f-2f42bfb37b5f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://86f9a77acae360740be827eb0d1e1516e8a49cb9b197b6f5f9468ac70a101158\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rk86l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:29Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-cvzg8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:39Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:39 crc kubenswrapper[4980]: I1206 03:34:39.283304 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cab608b3-a02f-475b-9aae-9b8f5d878de9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f7fcc44b42bcbb55509048e3a566bb557333c4a83e33d7b0fba8eb514317909\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c22e7dd8886c2dfbb742900ed86ec7b0c2edf9a1363e74f6666961d928f06c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://473e792b2d0a9644854ee0581c6120fc9c989bc342899ef72b1d056080fd4694\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ee8e936dfc39361d6f99460910591ca8c901a20f0d274bbb6aaebfa8a00ee06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://77bc12768b018625ba11a6bfd66688f926c4b7f3bc1d0e41c6551a18b485ecf3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://27cf07cd96893900da4afecd0d56fd9269bc60c1d23246b4685fcd8756357c19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://27cf07cd96893900da4afecd0d56fd9269bc60c1d23246b4685fcd8756357c19\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c49feb24566a8284378670fa0390710d38c237acc94ffb654f90f3fa4626e4c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c49feb24566a8284378670fa0390710d38c237acc94ffb654f90f3fa4626e4c8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:03Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://bb74d1c3d752a379fa394c450217badb45511fbcafad62f7b618131cab94b29e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bb74d1c3d752a379fa394c450217badb45511fbcafad62f7b618131cab94b29e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:01Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:39Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:39 crc kubenswrapper[4980]: I1206 03:34:39.295152 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/e6e88c5e-a255-449a-ae38-ac4d73a8e984-metrics-certs\") pod \"network-metrics-daemon-mcprh\" (UID: \"e6e88c5e-a255-449a-ae38-ac4d73a8e984\") " pod="openshift-multus/network-metrics-daemon-mcprh" Dec 06 03:34:39 crc kubenswrapper[4980]: E1206 03:34:39.295342 4980 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 06 03:34:39 crc kubenswrapper[4980]: E1206 03:34:39.295425 4980 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/e6e88c5e-a255-449a-ae38-ac4d73a8e984-metrics-certs podName:e6e88c5e-a255-449a-ae38-ac4d73a8e984 nodeName:}" failed. No retries permitted until 2025-12-06 03:34:40.295406664 +0000 UTC m=+39.540157935 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/e6e88c5e-a255-449a-ae38-ac4d73a8e984-metrics-certs") pod "network-metrics-daemon-mcprh" (UID: "e6e88c5e-a255-449a-ae38-ac4d73a8e984") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 06 03:34:39 crc kubenswrapper[4980]: I1206 03:34:39.300660 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://502c727d6baa31f400bee14c34b19abbe38af873f28f6a82a43547431755d981\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:39Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:39 crc kubenswrapper[4980]: I1206 03:34:39.306019 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:34:39 crc kubenswrapper[4980]: I1206 03:34:39.306052 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:34:39 crc kubenswrapper[4980]: I1206 03:34:39.306064 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:34:39 crc kubenswrapper[4980]: I1206 03:34:39.306084 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:34:39 crc kubenswrapper[4980]: I1206 03:34:39.306095 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:34:39Z","lastTransitionTime":"2025-12-06T03:34:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:34:39 crc kubenswrapper[4980]: I1206 03:34:39.317658 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-dx44q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5bac22a-512c-49d0-90db-47de6aaaefd0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://97469dea2cd5393da0acf784ccc1969c8a4d07da7d51b14cdc955013ee9710ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f7mdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3b5dab476c7f227a6e4a2eae8469432667fb9fad345d9419ba59feab9fd385e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c3b5dab476c7f227a6e4a2eae8469432667fb9fad345d9419ba59feab9fd385e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f7mdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e921499231cbdf26b8918c98f9ad8c322fd7848865caeef048470de3e2c0efba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e921499231cbdf26b8918c98f9ad8c322fd7848865caeef048470de3e2c0efba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f7mdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7cbfe7d8cdb89916c238296592fab45b9e92e2c585baa6d67230decd1c84c5bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7cbfe7d8cdb89916c238296592fab45b9e92e2c585baa6d67230decd1c84c5bb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f7mdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f45cbf23b6b0d208cefc4b493f901bab9d71ccf2f1456507d54fb1f30c4fd48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7f45cbf23b6b0d208cefc4b493f901bab9d71ccf2f1456507d54fb1f30c4fd48\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f7mdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://344ce4b145ff021fcaf163803d8ad8cc6e9565bf7f0992e2653c43db663d8411\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://344ce4b145ff021fcaf163803d8ad8cc6e9565bf7f0992e2653c43db663d8411\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f7mdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://619e56a9937dd6376ab471ba16da30e5d77f443f5bd33610bea5ec85cd0446ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://619e56a9937dd6376ab471ba16da30e5d77f443f5bd33610bea5ec85cd0446ef\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f7mdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:23Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-dx44q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:39Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:39 crc kubenswrapper[4980]: I1206 03:34:39.331157 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-r5zfb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"320f44d1-a671-4a91-b328-a8b0fdd8f23a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://92393ecaaef8cfcfd01837671f8cc89803dd570b637b4a373b40b7f487788826\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zhm5t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52e8db644463794041978ced2c0591c13e0ada8e5f1a78a94a0a665218a37b5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zhm5t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:23Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-r5zfb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:39Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:39 crc kubenswrapper[4980]: I1206 03:34:39.340471 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-mcprh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e6e88c5e-a255-449a-ae38-ac4d73a8e984\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g7xz9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g7xz9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:38Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-mcprh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:39Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:39 crc kubenswrapper[4980]: I1206 03:34:39.351559 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:39Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:39 crc kubenswrapper[4980]: I1206 03:34:39.361458 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4ef01e7d9809fd9cfc46e3a58b07159cdc6f5a3c210d1c9efc49bca27b30858\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:39Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:39 crc kubenswrapper[4980]: I1206 03:34:39.370649 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-kq4z8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f96c1f6c-55e8-4d86-bff3-5381581b7022\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://799498681afa7653408bc23b1e82d75c16e3c809d2ee7f322780e20c537eb710\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bdx7g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:23Z\\\"}}\" for pod \"openshift-multus\"/\"multus-kq4z8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:39Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:39 crc kubenswrapper[4980]: I1206 03:34:39.379188 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-6kxfq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"951b6deb-bcea-4332-8ae8-c23ac7aaef36\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:37Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:37Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b8htf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b8htf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:37Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-6kxfq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:39Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:39 crc kubenswrapper[4980]: I1206 03:34:39.389472 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7f958150-549d-4c80-8a7e-23c39e4f9500\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c28b6895b14e73f4f3eb7a8ce93076c05092ba9b58ea0e3e879c65bcfa227a21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://76cff8a2d652df38b04eefc053832baf4cbb5db778ff5e525b18c2825141b118\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://206f6f19fe3197f5d6798cc86c5473bf2decb02a2860ee7ffb5efb5f60952e8c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bafaa9670b05671fb57b658c2960bf3ee32c84e52f86a3a3bbab01d5a38ff0c4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://773a0bc2aab73ffc2717e987543defc9b786cc22300aa54b4ba4f3454504e331\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-06T03:34:22Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1206 03:34:15.305330 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1206 03:34:15.307249 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1592846546/tls.crt::/tmp/serving-cert-1592846546/tls.key\\\\\\\"\\\\nI1206 03:34:21.814037 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1206 03:34:21.816756 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1206 03:34:21.816780 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1206 03:34:21.816806 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1206 03:34:21.816813 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1206 03:34:21.831301 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1206 03:34:21.831311 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1206 03:34:21.831324 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 03:34:21.831330 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 03:34:21.831335 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1206 03:34:21.831339 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1206 03:34:21.831342 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1206 03:34:21.831346 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1206 03:34:21.911269 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:04Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d40e0e7446f8416bf9da94fc53c01be937f9a51a11eb101157d2ddd4597720d2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:04Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2cba89f2f1bed96d59f183628d900d9b72c91eb472e273c8bed0babcab485910\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2cba89f2f1bed96d59f183628d900d9b72c91eb472e273c8bed0babcab485910\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:01Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:39Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:39 crc kubenswrapper[4980]: I1206 03:34:39.406200 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6ed68f66-1cf6-465a-8f66-76d5af6a44d8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://725e84f6bdfb722a8ce69158d03ea4c66a9f087642d55cb9651e003c81be0c8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://71f7b492ea07cdf7dc65fc48184ec1e64914eb1811bbfdddc0fb14c8ad6cb296\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b543aa4ee03d5a14909b67d876c5675c02e1df2a89a3c9ac8d3fb587fea53d8f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://32af3427e872a70cb3c2357116f04cabc8d3d9b69e7306b2e0de2ae22f9469d4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:01Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:39Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:39 crc kubenswrapper[4980]: I1206 03:34:39.408032 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:34:39 crc kubenswrapper[4980]: I1206 03:34:39.408070 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:34:39 crc kubenswrapper[4980]: I1206 03:34:39.408079 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:34:39 crc kubenswrapper[4980]: I1206 03:34:39.408095 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:34:39 crc kubenswrapper[4980]: I1206 03:34:39.408105 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:34:39Z","lastTransitionTime":"2025-12-06T03:34:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:34:39 crc kubenswrapper[4980]: I1206 03:34:39.510872 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:34:39 crc kubenswrapper[4980]: I1206 03:34:39.510913 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:34:39 crc kubenswrapper[4980]: I1206 03:34:39.510922 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:34:39 crc kubenswrapper[4980]: I1206 03:34:39.510940 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:34:39 crc kubenswrapper[4980]: I1206 03:34:39.510952 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:34:39Z","lastTransitionTime":"2025-12-06T03:34:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:34:39 crc kubenswrapper[4980]: I1206 03:34:39.614410 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:34:39 crc kubenswrapper[4980]: I1206 03:34:39.614473 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:34:39 crc kubenswrapper[4980]: I1206 03:34:39.614490 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:34:39 crc kubenswrapper[4980]: I1206 03:34:39.614543 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:34:39 crc kubenswrapper[4980]: I1206 03:34:39.614564 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:34:39Z","lastTransitionTime":"2025-12-06T03:34:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:34:39 crc kubenswrapper[4980]: I1206 03:34:39.718589 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:34:39 crc kubenswrapper[4980]: I1206 03:34:39.718677 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:34:39 crc kubenswrapper[4980]: I1206 03:34:39.718689 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:34:39 crc kubenswrapper[4980]: I1206 03:34:39.718706 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:34:39 crc kubenswrapper[4980]: I1206 03:34:39.718717 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:34:39Z","lastTransitionTime":"2025-12-06T03:34:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:34:39 crc kubenswrapper[4980]: I1206 03:34:39.821427 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:34:39 crc kubenswrapper[4980]: I1206 03:34:39.821494 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:34:39 crc kubenswrapper[4980]: I1206 03:34:39.821554 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:34:39 crc kubenswrapper[4980]: I1206 03:34:39.821581 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:34:39 crc kubenswrapper[4980]: I1206 03:34:39.821599 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:34:39Z","lastTransitionTime":"2025-12-06T03:34:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:34:39 crc kubenswrapper[4980]: I1206 03:34:39.898424 4980 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-5w4fr_5af0e768-3c9c-4401-ab56-57516cd3170e/ovnkube-controller/1.log" Dec 06 03:34:39 crc kubenswrapper[4980]: I1206 03:34:39.899463 4980 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-5w4fr_5af0e768-3c9c-4401-ab56-57516cd3170e/ovnkube-controller/0.log" Dec 06 03:34:39 crc kubenswrapper[4980]: I1206 03:34:39.903281 4980 generic.go:334] "Generic (PLEG): container finished" podID="5af0e768-3c9c-4401-ab56-57516cd3170e" containerID="6cb5cc9f6f22133cebf095db595e5062d799d7e3c8ca0bf31d3809a296e507cc" exitCode=1 Dec 06 03:34:39 crc kubenswrapper[4980]: I1206 03:34:39.903418 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5w4fr" event={"ID":"5af0e768-3c9c-4401-ab56-57516cd3170e","Type":"ContainerDied","Data":"6cb5cc9f6f22133cebf095db595e5062d799d7e3c8ca0bf31d3809a296e507cc"} Dec 06 03:34:39 crc kubenswrapper[4980]: I1206 03:34:39.903557 4980 scope.go:117] "RemoveContainer" containerID="a8d3b4a48fbd8d6d81e7de4c07d3da18b7bfaa209461e10b38daf67000c1f699" Dec 06 03:34:39 crc kubenswrapper[4980]: I1206 03:34:39.904635 4980 scope.go:117] "RemoveContainer" containerID="6cb5cc9f6f22133cebf095db595e5062d799d7e3c8ca0bf31d3809a296e507cc" Dec 06 03:34:39 crc kubenswrapper[4980]: E1206 03:34:39.904964 4980 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-5w4fr_openshift-ovn-kubernetes(5af0e768-3c9c-4401-ab56-57516cd3170e)\"" pod="openshift-ovn-kubernetes/ovnkube-node-5w4fr" podUID="5af0e768-3c9c-4401-ab56-57516cd3170e" Dec 06 03:34:39 crc kubenswrapper[4980]: I1206 03:34:39.910209 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-6kxfq" event={"ID":"951b6deb-bcea-4332-8ae8-c23ac7aaef36","Type":"ContainerStarted","Data":"282df50dd9653c1cbe97fc2041d9e9ea91a992661682652aafc12e566d8f1b81"} Dec 06 03:34:39 crc kubenswrapper[4980]: I1206 03:34:39.925044 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:34:39 crc kubenswrapper[4980]: I1206 03:34:39.925105 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:34:39 crc kubenswrapper[4980]: I1206 03:34:39.925122 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:34:39 crc kubenswrapper[4980]: I1206 03:34:39.925142 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:34:39 crc kubenswrapper[4980]: I1206 03:34:39.925157 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:34:39Z","lastTransitionTime":"2025-12-06T03:34:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:34:39 crc kubenswrapper[4980]: I1206 03:34:39.930406 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7f958150-549d-4c80-8a7e-23c39e4f9500\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c28b6895b14e73f4f3eb7a8ce93076c05092ba9b58ea0e3e879c65bcfa227a21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://76cff8a2d652df38b04eefc053832baf4cbb5db778ff5e525b18c2825141b118\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://206f6f19fe3197f5d6798cc86c5473bf2decb02a2860ee7ffb5efb5f60952e8c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bafaa9670b05671fb57b658c2960bf3ee32c84e52f86a3a3bbab01d5a38ff0c4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://773a0bc2aab73ffc2717e987543defc9b786cc22300aa54b4ba4f3454504e331\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-06T03:34:22Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1206 03:34:15.305330 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1206 03:34:15.307249 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1592846546/tls.crt::/tmp/serving-cert-1592846546/tls.key\\\\\\\"\\\\nI1206 03:34:21.814037 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1206 03:34:21.816756 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1206 03:34:21.816780 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1206 03:34:21.816806 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1206 03:34:21.816813 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1206 03:34:21.831301 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1206 03:34:21.831311 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1206 03:34:21.831324 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 03:34:21.831330 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 03:34:21.831335 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1206 03:34:21.831339 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1206 03:34:21.831342 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1206 03:34:21.831346 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1206 03:34:21.911269 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:04Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d40e0e7446f8416bf9da94fc53c01be937f9a51a11eb101157d2ddd4597720d2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:04Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2cba89f2f1bed96d59f183628d900d9b72c91eb472e273c8bed0babcab485910\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2cba89f2f1bed96d59f183628d900d9b72c91eb472e273c8bed0babcab485910\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:01Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:39Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:39 crc kubenswrapper[4980]: I1206 03:34:39.949232 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6ed68f66-1cf6-465a-8f66-76d5af6a44d8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://725e84f6bdfb722a8ce69158d03ea4c66a9f087642d55cb9651e003c81be0c8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://71f7b492ea07cdf7dc65fc48184ec1e64914eb1811bbfdddc0fb14c8ad6cb296\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b543aa4ee03d5a14909b67d876c5675c02e1df2a89a3c9ac8d3fb587fea53d8f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://32af3427e872a70cb3c2357116f04cabc8d3d9b69e7306b2e0de2ae22f9469d4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:01Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:39Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:39 crc kubenswrapper[4980]: I1206 03:34:39.965218 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:22Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:39Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:39 crc kubenswrapper[4980]: I1206 03:34:39.976728 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-sv5sz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"873977aa-4eb0-4526-bc0b-757ad2dea2a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01f0cc855c822a71ea8ef6a7ce6e25747ff9956691735973f935c4a28672a7a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lg7tx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:23Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-sv5sz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:39Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:39 crc kubenswrapper[4980]: I1206 03:34:39.993574 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5w4fr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5af0e768-3c9c-4401-ab56-57516cd3170e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:23Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:23Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://768b8e1fe1a00a8467d744da5ad3f61006b0a3612c96f8e4d0ee6fcb93195f9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfzrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6342903b318317ee8d84c2f5d467cb25c811a789be0d6ab0e0cba37b5611e16d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfzrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d620685574db7594132ae24a91ec0a93dce20a9a289540deebb39650b8a1816\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfzrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c67ec8f9c1764fcdf215d52822c2d4f0437dc7f056f05d708767121f6c1f3515\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfzrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f35beae5339eeaf26494ed678e3c1bd02349eae2a4b4615325b17923f3b653b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfzrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a940ec6ee523313728f7e8ae934b7e0b0cff82771e3e64723309db1fae013c3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfzrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6cb5cc9f6f22133cebf095db595e5062d799d7e3c8ca0bf31d3809a296e507cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a8d3b4a48fbd8d6d81e7de4c07d3da18b7bfaa209461e10b38daf67000c1f699\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-06T03:34:37Z\\\",\\\"message\\\":\\\":default/a8519615025667110816) with []\\\\nI1206 03:34:37.270920 6168 address_set.go:302] New(bf133528-8652-4c84-85ff-881f0afe9837/default-network-controller:EgressService:egresssvc-served-pods:v4/a13607449821398607916) with []\\\\nI1206 03:34:37.270987 6168 factory.go:1336] Added *v1.Node event handler 7\\\\nI1206 03:34:37.271037 6168 factory.go:1336] Added *v1.EgressIP event handler 8\\\\nI1206 03:34:37.271038 6168 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1206 03:34:37.271088 6168 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1206 03:34:37.271123 6168 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1206 03:34:37.271230 6168 factory.go:656] Stopping watch factory\\\\nI1206 03:34:37.271122 6168 handler.go:208] Removed *v1.Node event handler 2\\\\nI1206 03:34:37.271267 6168 handler.go:208] Removed *v1.Node event handler 7\\\\nI1206 03:34:37.271295 6168 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1206 03:34:37.271581 6168 factory.go:1336] Added *v1.EgressFirewall event handler 9\\\\nI1206 03:34:37.271672 6168 controller.go:132] Adding controller ef_node_controller event handlers\\\\nI1206 03:34:37.271709 6168 ovnkube.go:599] Stopped ovnkube\\\\nI1206 03:34:37.271735 6168 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1206 03:34:37.271795 6168 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:32Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6cb5cc9f6f22133cebf095db595e5062d799d7e3c8ca0bf31d3809a296e507cc\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-06T03:34:39Z\\\",\\\"message\\\":\\\"lector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1206 03:34:39.306579 6370 reflector.go:311] Stopping reflector *v1.AdminPolicyBasedExternalRoute (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/adminpolicybasedroute/v1/apis/informers/externalversions/factory.go:140\\\\nI1206 03:34:39.306665 6370 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1206 03:34:39.306723 6370 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI1206 03:34:39.306893 6370 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1206 03:34:39.307227 6370 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1206 03:34:39.307306 6370 reflector.go:311] Stopping reflector *v1.NetworkAttachmentDefinition (0s) from github.com/k8snetworkplumbingwg/network-attachment-definition-client/pkg/client/informers/externalversions/factory.go:117\\\\nI1206 03:34:39.307466 6370 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfzrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe261f4322b85702181d7473ab2a9a4d2d4b869d586c3091f811b12a81a63aae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfzrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d876df04b5ab0ba07a5cbcafbe4648403c5014c59eb7d40adee2657c3a255473\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d876df04b5ab0ba07a5cbcafbe4648403c5014c59eb7d40adee2657c3a255473\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfzrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:23Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-5w4fr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:39Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:40 crc kubenswrapper[4980]: I1206 03:34:40.013416 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:22Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:40Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:40 crc kubenswrapper[4980]: I1206 03:34:40.028179 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://334ea6a273fd7d268906c9f4a6dbf6f4df4fcdbf38f3d7d248cb7d6cf8537794\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ec44dcd5e4bf0a0466bb586ebd61259364dbbfe22499f4cbff854e3bd153082\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:40Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:40 crc kubenswrapper[4980]: I1206 03:34:40.028280 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:34:40 crc kubenswrapper[4980]: I1206 03:34:40.028333 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:34:40 crc kubenswrapper[4980]: I1206 03:34:40.028352 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:34:40 crc kubenswrapper[4980]: I1206 03:34:40.028377 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:34:40 crc kubenswrapper[4980]: I1206 03:34:40.028395 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:34:40Z","lastTransitionTime":"2025-12-06T03:34:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:34:40 crc kubenswrapper[4980]: I1206 03:34:40.041236 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-r5zfb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"320f44d1-a671-4a91-b328-a8b0fdd8f23a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://92393ecaaef8cfcfd01837671f8cc89803dd570b637b4a373b40b7f487788826\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zhm5t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52e8db644463794041978ced2c0591c13e0ada8e5f1a78a94a0a665218a37b5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zhm5t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:23Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-r5zfb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:40Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:40 crc kubenswrapper[4980]: I1206 03:34:40.056383 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-cvzg8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f1cde706-268d-4d26-818f-2f42bfb37b5f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://86f9a77acae360740be827eb0d1e1516e8a49cb9b197b6f5f9468ac70a101158\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rk86l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:29Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-cvzg8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:40Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:40 crc kubenswrapper[4980]: I1206 03:34:40.087109 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cab608b3-a02f-475b-9aae-9b8f5d878de9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f7fcc44b42bcbb55509048e3a566bb557333c4a83e33d7b0fba8eb514317909\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c22e7dd8886c2dfbb742900ed86ec7b0c2edf9a1363e74f6666961d928f06c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://473e792b2d0a9644854ee0581c6120fc9c989bc342899ef72b1d056080fd4694\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ee8e936dfc39361d6f99460910591ca8c901a20f0d274bbb6aaebfa8a00ee06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://77bc12768b018625ba11a6bfd66688f926c4b7f3bc1d0e41c6551a18b485ecf3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://27cf07cd96893900da4afecd0d56fd9269bc60c1d23246b4685fcd8756357c19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://27cf07cd96893900da4afecd0d56fd9269bc60c1d23246b4685fcd8756357c19\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c49feb24566a8284378670fa0390710d38c237acc94ffb654f90f3fa4626e4c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c49feb24566a8284378670fa0390710d38c237acc94ffb654f90f3fa4626e4c8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:03Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://bb74d1c3d752a379fa394c450217badb45511fbcafad62f7b618131cab94b29e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bb74d1c3d752a379fa394c450217badb45511fbcafad62f7b618131cab94b29e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:01Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:40Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:40 crc kubenswrapper[4980]: I1206 03:34:40.107317 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://502c727d6baa31f400bee14c34b19abbe38af873f28f6a82a43547431755d981\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:40Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:40 crc kubenswrapper[4980]: I1206 03:34:40.130012 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-dx44q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5bac22a-512c-49d0-90db-47de6aaaefd0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://97469dea2cd5393da0acf784ccc1969c8a4d07da7d51b14cdc955013ee9710ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f7mdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3b5dab476c7f227a6e4a2eae8469432667fb9fad345d9419ba59feab9fd385e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c3b5dab476c7f227a6e4a2eae8469432667fb9fad345d9419ba59feab9fd385e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f7mdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e921499231cbdf26b8918c98f9ad8c322fd7848865caeef048470de3e2c0efba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e921499231cbdf26b8918c98f9ad8c322fd7848865caeef048470de3e2c0efba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f7mdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7cbfe7d8cdb89916c238296592fab45b9e92e2c585baa6d67230decd1c84c5bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7cbfe7d8cdb89916c238296592fab45b9e92e2c585baa6d67230decd1c84c5bb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f7mdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f45cbf23b6b0d208cefc4b493f901bab9d71ccf2f1456507d54fb1f30c4fd48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7f45cbf23b6b0d208cefc4b493f901bab9d71ccf2f1456507d54fb1f30c4fd48\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f7mdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://344ce4b145ff021fcaf163803d8ad8cc6e9565bf7f0992e2653c43db663d8411\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://344ce4b145ff021fcaf163803d8ad8cc6e9565bf7f0992e2653c43db663d8411\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f7mdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://619e56a9937dd6376ab471ba16da30e5d77f443f5bd33610bea5ec85cd0446ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://619e56a9937dd6376ab471ba16da30e5d77f443f5bd33610bea5ec85cd0446ef\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f7mdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:23Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-dx44q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:40Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:40 crc kubenswrapper[4980]: I1206 03:34:40.130866 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:34:40 crc kubenswrapper[4980]: I1206 03:34:40.130920 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:34:40 crc kubenswrapper[4980]: I1206 03:34:40.130934 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:34:40 crc kubenswrapper[4980]: I1206 03:34:40.130958 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:34:40 crc kubenswrapper[4980]: I1206 03:34:40.130974 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:34:40Z","lastTransitionTime":"2025-12-06T03:34:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:34:40 crc kubenswrapper[4980]: I1206 03:34:40.142143 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-6kxfq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"951b6deb-bcea-4332-8ae8-c23ac7aaef36\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:37Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:37Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b8htf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b8htf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:37Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-6kxfq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:40Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:40 crc kubenswrapper[4980]: I1206 03:34:40.153929 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-mcprh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e6e88c5e-a255-449a-ae38-ac4d73a8e984\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g7xz9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g7xz9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:38Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-mcprh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:40Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:40 crc kubenswrapper[4980]: I1206 03:34:40.166051 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:40Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:40 crc kubenswrapper[4980]: I1206 03:34:40.182747 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4ef01e7d9809fd9cfc46e3a58b07159cdc6f5a3c210d1c9efc49bca27b30858\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:40Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:40 crc kubenswrapper[4980]: I1206 03:34:40.183936 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 03:34:40 crc kubenswrapper[4980]: I1206 03:34:40.183972 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 03:34:40 crc kubenswrapper[4980]: I1206 03:34:40.183950 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-mcprh" Dec 06 03:34:40 crc kubenswrapper[4980]: E1206 03:34:40.184098 4980 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 03:34:40 crc kubenswrapper[4980]: E1206 03:34:40.184168 4980 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-mcprh" podUID="e6e88c5e-a255-449a-ae38-ac4d73a8e984" Dec 06 03:34:40 crc kubenswrapper[4980]: E1206 03:34:40.184220 4980 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 03:34:40 crc kubenswrapper[4980]: I1206 03:34:40.199702 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-kq4z8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f96c1f6c-55e8-4d86-bff3-5381581b7022\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://799498681afa7653408bc23b1e82d75c16e3c809d2ee7f322780e20c537eb710\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bdx7g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:23Z\\\"}}\" for pod \"openshift-multus\"/\"multus-kq4z8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:40Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:40 crc kubenswrapper[4980]: I1206 03:34:40.217856 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:22Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:40Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:40 crc kubenswrapper[4980]: I1206 03:34:40.234058 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:34:40 crc kubenswrapper[4980]: I1206 03:34:40.234122 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:34:40 crc kubenswrapper[4980]: I1206 03:34:40.234132 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:34:40 crc kubenswrapper[4980]: I1206 03:34:40.234166 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:34:40 crc kubenswrapper[4980]: I1206 03:34:40.234178 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:34:40Z","lastTransitionTime":"2025-12-06T03:34:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:34:40 crc kubenswrapper[4980]: I1206 03:34:40.237240 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://334ea6a273fd7d268906c9f4a6dbf6f4df4fcdbf38f3d7d248cb7d6cf8537794\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ec44dcd5e4bf0a0466bb586ebd61259364dbbfe22499f4cbff854e3bd153082\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:40Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:40 crc kubenswrapper[4980]: I1206 03:34:40.255899 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:22Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:40Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:40 crc kubenswrapper[4980]: I1206 03:34:40.269845 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-sv5sz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"873977aa-4eb0-4526-bc0b-757ad2dea2a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01f0cc855c822a71ea8ef6a7ce6e25747ff9956691735973f935c4a28672a7a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lg7tx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:23Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-sv5sz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:40Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:40 crc kubenswrapper[4980]: I1206 03:34:40.299836 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5w4fr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5af0e768-3c9c-4401-ab56-57516cd3170e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:23Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:23Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://768b8e1fe1a00a8467d744da5ad3f61006b0a3612c96f8e4d0ee6fcb93195f9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfzrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6342903b318317ee8d84c2f5d467cb25c811a789be0d6ab0e0cba37b5611e16d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfzrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d620685574db7594132ae24a91ec0a93dce20a9a289540deebb39650b8a1816\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfzrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c67ec8f9c1764fcdf215d52822c2d4f0437dc7f056f05d708767121f6c1f3515\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfzrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f35beae5339eeaf26494ed678e3c1bd02349eae2a4b4615325b17923f3b653b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfzrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a940ec6ee523313728f7e8ae934b7e0b0cff82771e3e64723309db1fae013c3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfzrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6cb5cc9f6f22133cebf095db595e5062d799d7e3c8ca0bf31d3809a296e507cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a8d3b4a48fbd8d6d81e7de4c07d3da18b7bfaa209461e10b38daf67000c1f699\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-06T03:34:37Z\\\",\\\"message\\\":\\\":default/a8519615025667110816) with []\\\\nI1206 03:34:37.270920 6168 address_set.go:302] New(bf133528-8652-4c84-85ff-881f0afe9837/default-network-controller:EgressService:egresssvc-served-pods:v4/a13607449821398607916) with []\\\\nI1206 03:34:37.270987 6168 factory.go:1336] Added *v1.Node event handler 7\\\\nI1206 03:34:37.271037 6168 factory.go:1336] Added *v1.EgressIP event handler 8\\\\nI1206 03:34:37.271038 6168 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1206 03:34:37.271088 6168 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1206 03:34:37.271123 6168 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1206 03:34:37.271230 6168 factory.go:656] Stopping watch factory\\\\nI1206 03:34:37.271122 6168 handler.go:208] Removed *v1.Node event handler 2\\\\nI1206 03:34:37.271267 6168 handler.go:208] Removed *v1.Node event handler 7\\\\nI1206 03:34:37.271295 6168 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1206 03:34:37.271581 6168 factory.go:1336] Added *v1.EgressFirewall event handler 9\\\\nI1206 03:34:37.271672 6168 controller.go:132] Adding controller ef_node_controller event handlers\\\\nI1206 03:34:37.271709 6168 ovnkube.go:599] Stopped ovnkube\\\\nI1206 03:34:37.271735 6168 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1206 03:34:37.271795 6168 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:32Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6cb5cc9f6f22133cebf095db595e5062d799d7e3c8ca0bf31d3809a296e507cc\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-06T03:34:39Z\\\",\\\"message\\\":\\\"lector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1206 03:34:39.306579 6370 reflector.go:311] Stopping reflector *v1.AdminPolicyBasedExternalRoute (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/adminpolicybasedroute/v1/apis/informers/externalversions/factory.go:140\\\\nI1206 03:34:39.306665 6370 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1206 03:34:39.306723 6370 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI1206 03:34:39.306893 6370 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1206 03:34:39.307227 6370 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1206 03:34:39.307306 6370 reflector.go:311] Stopping reflector *v1.NetworkAttachmentDefinition (0s) from github.com/k8snetworkplumbingwg/network-attachment-definition-client/pkg/client/informers/externalversions/factory.go:117\\\\nI1206 03:34:39.307466 6370 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfzrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe261f4322b85702181d7473ab2a9a4d2d4b869d586c3091f811b12a81a63aae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfzrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d876df04b5ab0ba07a5cbcafbe4648403c5014c59eb7d40adee2657c3a255473\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d876df04b5ab0ba07a5cbcafbe4648403c5014c59eb7d40adee2657c3a255473\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfzrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:23Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-5w4fr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:40Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:40 crc kubenswrapper[4980]: E1206 03:34:40.304224 4980 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 06 03:34:40 crc kubenswrapper[4980]: I1206 03:34:40.304210 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/e6e88c5e-a255-449a-ae38-ac4d73a8e984-metrics-certs\") pod \"network-metrics-daemon-mcprh\" (UID: \"e6e88c5e-a255-449a-ae38-ac4d73a8e984\") " pod="openshift-multus/network-metrics-daemon-mcprh" Dec 06 03:34:40 crc kubenswrapper[4980]: E1206 03:34:40.304294 4980 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/e6e88c5e-a255-449a-ae38-ac4d73a8e984-metrics-certs podName:e6e88c5e-a255-449a-ae38-ac4d73a8e984 nodeName:}" failed. No retries permitted until 2025-12-06 03:34:42.304275903 +0000 UTC m=+41.549027194 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/e6e88c5e-a255-449a-ae38-ac4d73a8e984-metrics-certs") pod "network-metrics-daemon-mcprh" (UID: "e6e88c5e-a255-449a-ae38-ac4d73a8e984") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 06 03:34:40 crc kubenswrapper[4980]: I1206 03:34:40.329819 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cab608b3-a02f-475b-9aae-9b8f5d878de9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f7fcc44b42bcbb55509048e3a566bb557333c4a83e33d7b0fba8eb514317909\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c22e7dd8886c2dfbb742900ed86ec7b0c2edf9a1363e74f6666961d928f06c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://473e792b2d0a9644854ee0581c6120fc9c989bc342899ef72b1d056080fd4694\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ee8e936dfc39361d6f99460910591ca8c901a20f0d274bbb6aaebfa8a00ee06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://77bc12768b018625ba11a6bfd66688f926c4b7f3bc1d0e41c6551a18b485ecf3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://27cf07cd96893900da4afecd0d56fd9269bc60c1d23246b4685fcd8756357c19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://27cf07cd96893900da4afecd0d56fd9269bc60c1d23246b4685fcd8756357c19\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c49feb24566a8284378670fa0390710d38c237acc94ffb654f90f3fa4626e4c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c49feb24566a8284378670fa0390710d38c237acc94ffb654f90f3fa4626e4c8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:03Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://bb74d1c3d752a379fa394c450217badb45511fbcafad62f7b618131cab94b29e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bb74d1c3d752a379fa394c450217badb45511fbcafad62f7b618131cab94b29e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:01Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:40Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:40 crc kubenswrapper[4980]: I1206 03:34:40.337365 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:34:40 crc kubenswrapper[4980]: I1206 03:34:40.337417 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:34:40 crc kubenswrapper[4980]: I1206 03:34:40.337430 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:34:40 crc kubenswrapper[4980]: I1206 03:34:40.337448 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:34:40 crc kubenswrapper[4980]: I1206 03:34:40.337459 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:34:40Z","lastTransitionTime":"2025-12-06T03:34:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:34:40 crc kubenswrapper[4980]: I1206 03:34:40.370083 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://502c727d6baa31f400bee14c34b19abbe38af873f28f6a82a43547431755d981\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:40Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:40 crc kubenswrapper[4980]: I1206 03:34:40.424242 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-dx44q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5bac22a-512c-49d0-90db-47de6aaaefd0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://97469dea2cd5393da0acf784ccc1969c8a4d07da7d51b14cdc955013ee9710ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f7mdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3b5dab476c7f227a6e4a2eae8469432667fb9fad345d9419ba59feab9fd385e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c3b5dab476c7f227a6e4a2eae8469432667fb9fad345d9419ba59feab9fd385e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f7mdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e921499231cbdf26b8918c98f9ad8c322fd7848865caeef048470de3e2c0efba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e921499231cbdf26b8918c98f9ad8c322fd7848865caeef048470de3e2c0efba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f7mdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7cbfe7d8cdb89916c238296592fab45b9e92e2c585baa6d67230decd1c84c5bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7cbfe7d8cdb89916c238296592fab45b9e92e2c585baa6d67230decd1c84c5bb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f7mdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f45cbf23b6b0d208cefc4b493f901bab9d71ccf2f1456507d54fb1f30c4fd48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7f45cbf23b6b0d208cefc4b493f901bab9d71ccf2f1456507d54fb1f30c4fd48\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f7mdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://344ce4b145ff021fcaf163803d8ad8cc6e9565bf7f0992e2653c43db663d8411\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://344ce4b145ff021fcaf163803d8ad8cc6e9565bf7f0992e2653c43db663d8411\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f7mdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://619e56a9937dd6376ab471ba16da30e5d77f443f5bd33610bea5ec85cd0446ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://619e56a9937dd6376ab471ba16da30e5d77f443f5bd33610bea5ec85cd0446ef\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f7mdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:23Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-dx44q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:40Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:40 crc kubenswrapper[4980]: I1206 03:34:40.440984 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:34:40 crc kubenswrapper[4980]: I1206 03:34:40.441048 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:34:40 crc kubenswrapper[4980]: I1206 03:34:40.441070 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:34:40 crc kubenswrapper[4980]: I1206 03:34:40.441102 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:34:40 crc kubenswrapper[4980]: I1206 03:34:40.441123 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:34:40Z","lastTransitionTime":"2025-12-06T03:34:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:34:40 crc kubenswrapper[4980]: I1206 03:34:40.458608 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-r5zfb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"320f44d1-a671-4a91-b328-a8b0fdd8f23a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://92393ecaaef8cfcfd01837671f8cc89803dd570b637b4a373b40b7f487788826\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zhm5t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52e8db644463794041978ced2c0591c13e0ada8e5f1a78a94a0a665218a37b5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zhm5t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:23Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-r5zfb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:40Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:40 crc kubenswrapper[4980]: I1206 03:34:40.482030 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-cvzg8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f1cde706-268d-4d26-818f-2f42bfb37b5f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://86f9a77acae360740be827eb0d1e1516e8a49cb9b197b6f5f9468ac70a101158\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rk86l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:29Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-cvzg8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:40Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:40 crc kubenswrapper[4980]: I1206 03:34:40.523829 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:40Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:40 crc kubenswrapper[4980]: I1206 03:34:40.544079 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:34:40 crc kubenswrapper[4980]: I1206 03:34:40.544124 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:34:40 crc kubenswrapper[4980]: I1206 03:34:40.544146 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:34:40 crc kubenswrapper[4980]: I1206 03:34:40.544163 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:34:40 crc kubenswrapper[4980]: I1206 03:34:40.544174 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:34:40Z","lastTransitionTime":"2025-12-06T03:34:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:34:40 crc kubenswrapper[4980]: I1206 03:34:40.572884 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4ef01e7d9809fd9cfc46e3a58b07159cdc6f5a3c210d1c9efc49bca27b30858\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:40Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:40 crc kubenswrapper[4980]: I1206 03:34:40.610635 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-kq4z8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f96c1f6c-55e8-4d86-bff3-5381581b7022\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://799498681afa7653408bc23b1e82d75c16e3c809d2ee7f322780e20c537eb710\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bdx7g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:23Z\\\"}}\" for pod \"openshift-multus\"/\"multus-kq4z8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:40Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:40 crc kubenswrapper[4980]: I1206 03:34:40.646363 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-6kxfq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"951b6deb-bcea-4332-8ae8-c23ac7aaef36\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://423d30c3e65b04bdf6ff5ef53d8a7575a434c05e28489370a7370381ebc29de9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b8htf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://282df50dd9653c1cbe97fc2041d9e9ea91a992661682652aafc12e566d8f1b81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b8htf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:37Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-6kxfq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:40Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:40 crc kubenswrapper[4980]: I1206 03:34:40.647724 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:34:40 crc kubenswrapper[4980]: I1206 03:34:40.648016 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:34:40 crc kubenswrapper[4980]: I1206 03:34:40.648156 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:34:40 crc kubenswrapper[4980]: I1206 03:34:40.648333 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:34:40 crc kubenswrapper[4980]: I1206 03:34:40.648471 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:34:40Z","lastTransitionTime":"2025-12-06T03:34:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:34:40 crc kubenswrapper[4980]: I1206 03:34:40.684215 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-mcprh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e6e88c5e-a255-449a-ae38-ac4d73a8e984\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g7xz9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g7xz9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:38Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-mcprh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:40Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:40 crc kubenswrapper[4980]: I1206 03:34:40.731499 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7f958150-549d-4c80-8a7e-23c39e4f9500\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c28b6895b14e73f4f3eb7a8ce93076c05092ba9b58ea0e3e879c65bcfa227a21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://76cff8a2d652df38b04eefc053832baf4cbb5db778ff5e525b18c2825141b118\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://206f6f19fe3197f5d6798cc86c5473bf2decb02a2860ee7ffb5efb5f60952e8c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bafaa9670b05671fb57b658c2960bf3ee32c84e52f86a3a3bbab01d5a38ff0c4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://773a0bc2aab73ffc2717e987543defc9b786cc22300aa54b4ba4f3454504e331\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-06T03:34:22Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1206 03:34:15.305330 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1206 03:34:15.307249 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1592846546/tls.crt::/tmp/serving-cert-1592846546/tls.key\\\\\\\"\\\\nI1206 03:34:21.814037 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1206 03:34:21.816756 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1206 03:34:21.816780 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1206 03:34:21.816806 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1206 03:34:21.816813 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1206 03:34:21.831301 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1206 03:34:21.831311 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1206 03:34:21.831324 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 03:34:21.831330 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 03:34:21.831335 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1206 03:34:21.831339 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1206 03:34:21.831342 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1206 03:34:21.831346 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1206 03:34:21.911269 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:04Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d40e0e7446f8416bf9da94fc53c01be937f9a51a11eb101157d2ddd4597720d2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:04Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2cba89f2f1bed96d59f183628d900d9b72c91eb472e273c8bed0babcab485910\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2cba89f2f1bed96d59f183628d900d9b72c91eb472e273c8bed0babcab485910\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:01Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:40Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:40 crc kubenswrapper[4980]: I1206 03:34:40.751443 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:34:40 crc kubenswrapper[4980]: I1206 03:34:40.751558 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:34:40 crc kubenswrapper[4980]: I1206 03:34:40.751578 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:34:40 crc kubenswrapper[4980]: I1206 03:34:40.751609 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:34:40 crc kubenswrapper[4980]: I1206 03:34:40.751634 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:34:40Z","lastTransitionTime":"2025-12-06T03:34:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:34:40 crc kubenswrapper[4980]: I1206 03:34:40.769280 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6ed68f66-1cf6-465a-8f66-76d5af6a44d8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://725e84f6bdfb722a8ce69158d03ea4c66a9f087642d55cb9651e003c81be0c8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://71f7b492ea07cdf7dc65fc48184ec1e64914eb1811bbfdddc0fb14c8ad6cb296\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b543aa4ee03d5a14909b67d876c5675c02e1df2a89a3c9ac8d3fb587fea53d8f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://32af3427e872a70cb3c2357116f04cabc8d3d9b69e7306b2e0de2ae22f9469d4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:01Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:40Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:40 crc kubenswrapper[4980]: I1206 03:34:40.855485 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:34:40 crc kubenswrapper[4980]: I1206 03:34:40.855579 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:34:40 crc kubenswrapper[4980]: I1206 03:34:40.855603 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:34:40 crc kubenswrapper[4980]: I1206 03:34:40.855632 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:34:40 crc kubenswrapper[4980]: I1206 03:34:40.855652 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:34:40Z","lastTransitionTime":"2025-12-06T03:34:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:34:40 crc kubenswrapper[4980]: I1206 03:34:40.918052 4980 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-5w4fr_5af0e768-3c9c-4401-ab56-57516cd3170e/ovnkube-controller/1.log" Dec 06 03:34:40 crc kubenswrapper[4980]: I1206 03:34:40.923499 4980 scope.go:117] "RemoveContainer" containerID="6cb5cc9f6f22133cebf095db595e5062d799d7e3c8ca0bf31d3809a296e507cc" Dec 06 03:34:40 crc kubenswrapper[4980]: E1206 03:34:40.923779 4980 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-5w4fr_openshift-ovn-kubernetes(5af0e768-3c9c-4401-ab56-57516cd3170e)\"" pod="openshift-ovn-kubernetes/ovnkube-node-5w4fr" podUID="5af0e768-3c9c-4401-ab56-57516cd3170e" Dec 06 03:34:40 crc kubenswrapper[4980]: I1206 03:34:40.940835 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:40Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:40 crc kubenswrapper[4980]: I1206 03:34:40.958268 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:34:40 crc kubenswrapper[4980]: I1206 03:34:40.958354 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:34:40 crc kubenswrapper[4980]: I1206 03:34:40.958378 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:34:40 crc kubenswrapper[4980]: I1206 03:34:40.958408 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:34:40 crc kubenswrapper[4980]: I1206 03:34:40.958431 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:34:40Z","lastTransitionTime":"2025-12-06T03:34:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:34:40 crc kubenswrapper[4980]: I1206 03:34:40.960407 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4ef01e7d9809fd9cfc46e3a58b07159cdc6f5a3c210d1c9efc49bca27b30858\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:40Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:40 crc kubenswrapper[4980]: I1206 03:34:40.980902 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-kq4z8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f96c1f6c-55e8-4d86-bff3-5381581b7022\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://799498681afa7653408bc23b1e82d75c16e3c809d2ee7f322780e20c537eb710\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bdx7g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:23Z\\\"}}\" for pod \"openshift-multus\"/\"multus-kq4z8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:40Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:40 crc kubenswrapper[4980]: I1206 03:34:40.996212 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-6kxfq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"951b6deb-bcea-4332-8ae8-c23ac7aaef36\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://423d30c3e65b04bdf6ff5ef53d8a7575a434c05e28489370a7370381ebc29de9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b8htf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://282df50dd9653c1cbe97fc2041d9e9ea91a992661682652aafc12e566d8f1b81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b8htf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:37Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-6kxfq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:40Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:41 crc kubenswrapper[4980]: I1206 03:34:41.009851 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-mcprh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e6e88c5e-a255-449a-ae38-ac4d73a8e984\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g7xz9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g7xz9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:38Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-mcprh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:41Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:41 crc kubenswrapper[4980]: I1206 03:34:41.035961 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7f958150-549d-4c80-8a7e-23c39e4f9500\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c28b6895b14e73f4f3eb7a8ce93076c05092ba9b58ea0e3e879c65bcfa227a21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://76cff8a2d652df38b04eefc053832baf4cbb5db778ff5e525b18c2825141b118\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://206f6f19fe3197f5d6798cc86c5473bf2decb02a2860ee7ffb5efb5f60952e8c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bafaa9670b05671fb57b658c2960bf3ee32c84e52f86a3a3bbab01d5a38ff0c4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://773a0bc2aab73ffc2717e987543defc9b786cc22300aa54b4ba4f3454504e331\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-06T03:34:22Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1206 03:34:15.305330 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1206 03:34:15.307249 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1592846546/tls.crt::/tmp/serving-cert-1592846546/tls.key\\\\\\\"\\\\nI1206 03:34:21.814037 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1206 03:34:21.816756 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1206 03:34:21.816780 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1206 03:34:21.816806 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1206 03:34:21.816813 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1206 03:34:21.831301 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1206 03:34:21.831311 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1206 03:34:21.831324 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 03:34:21.831330 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 03:34:21.831335 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1206 03:34:21.831339 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1206 03:34:21.831342 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1206 03:34:21.831346 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1206 03:34:21.911269 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:04Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d40e0e7446f8416bf9da94fc53c01be937f9a51a11eb101157d2ddd4597720d2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:04Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2cba89f2f1bed96d59f183628d900d9b72c91eb472e273c8bed0babcab485910\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2cba89f2f1bed96d59f183628d900d9b72c91eb472e273c8bed0babcab485910\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:01Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:41Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:41 crc kubenswrapper[4980]: I1206 03:34:41.051917 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6ed68f66-1cf6-465a-8f66-76d5af6a44d8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://725e84f6bdfb722a8ce69158d03ea4c66a9f087642d55cb9651e003c81be0c8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://71f7b492ea07cdf7dc65fc48184ec1e64914eb1811bbfdddc0fb14c8ad6cb296\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b543aa4ee03d5a14909b67d876c5675c02e1df2a89a3c9ac8d3fb587fea53d8f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://32af3427e872a70cb3c2357116f04cabc8d3d9b69e7306b2e0de2ae22f9469d4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:01Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:41Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:41 crc kubenswrapper[4980]: I1206 03:34:41.062892 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:34:41 crc kubenswrapper[4980]: I1206 03:34:41.062952 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:34:41 crc kubenswrapper[4980]: I1206 03:34:41.062968 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:34:41 crc kubenswrapper[4980]: I1206 03:34:41.062993 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:34:41 crc kubenswrapper[4980]: I1206 03:34:41.063009 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:34:41Z","lastTransitionTime":"2025-12-06T03:34:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:34:41 crc kubenswrapper[4980]: I1206 03:34:41.088097 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:22Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:41Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:41 crc kubenswrapper[4980]: I1206 03:34:41.127695 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://334ea6a273fd7d268906c9f4a6dbf6f4df4fcdbf38f3d7d248cb7d6cf8537794\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ec44dcd5e4bf0a0466bb586ebd61259364dbbfe22499f4cbff854e3bd153082\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:41Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:41 crc kubenswrapper[4980]: I1206 03:34:41.162850 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:22Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:41Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:41 crc kubenswrapper[4980]: I1206 03:34:41.165689 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:34:41 crc kubenswrapper[4980]: I1206 03:34:41.165855 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:34:41 crc kubenswrapper[4980]: I1206 03:34:41.165958 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:34:41 crc kubenswrapper[4980]: I1206 03:34:41.166057 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:34:41 crc kubenswrapper[4980]: I1206 03:34:41.166159 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:34:41Z","lastTransitionTime":"2025-12-06T03:34:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:34:41 crc kubenswrapper[4980]: I1206 03:34:41.183647 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 03:34:41 crc kubenswrapper[4980]: E1206 03:34:41.184033 4980 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 03:34:41 crc kubenswrapper[4980]: I1206 03:34:41.203264 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-sv5sz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"873977aa-4eb0-4526-bc0b-757ad2dea2a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01f0cc855c822a71ea8ef6a7ce6e25747ff9956691735973f935c4a28672a7a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lg7tx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:23Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-sv5sz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:41Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:41 crc kubenswrapper[4980]: I1206 03:34:41.259113 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5w4fr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5af0e768-3c9c-4401-ab56-57516cd3170e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:23Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:23Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://768b8e1fe1a00a8467d744da5ad3f61006b0a3612c96f8e4d0ee6fcb93195f9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfzrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6342903b318317ee8d84c2f5d467cb25c811a789be0d6ab0e0cba37b5611e16d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfzrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d620685574db7594132ae24a91ec0a93dce20a9a289540deebb39650b8a1816\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfzrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c67ec8f9c1764fcdf215d52822c2d4f0437dc7f056f05d708767121f6c1f3515\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfzrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f35beae5339eeaf26494ed678e3c1bd02349eae2a4b4615325b17923f3b653b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfzrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a940ec6ee523313728f7e8ae934b7e0b0cff82771e3e64723309db1fae013c3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfzrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6cb5cc9f6f22133cebf095db595e5062d799d7e3c8ca0bf31d3809a296e507cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6cb5cc9f6f22133cebf095db595e5062d799d7e3c8ca0bf31d3809a296e507cc\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-06T03:34:39Z\\\",\\\"message\\\":\\\"lector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1206 03:34:39.306579 6370 reflector.go:311] Stopping reflector *v1.AdminPolicyBasedExternalRoute (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/adminpolicybasedroute/v1/apis/informers/externalversions/factory.go:140\\\\nI1206 03:34:39.306665 6370 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1206 03:34:39.306723 6370 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI1206 03:34:39.306893 6370 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1206 03:34:39.307227 6370 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1206 03:34:39.307306 6370 reflector.go:311] Stopping reflector *v1.NetworkAttachmentDefinition (0s) from github.com/k8snetworkplumbingwg/network-attachment-definition-client/pkg/client/informers/externalversions/factory.go:117\\\\nI1206 03:34:39.307466 6370 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:37Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-5w4fr_openshift-ovn-kubernetes(5af0e768-3c9c-4401-ab56-57516cd3170e)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfzrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe261f4322b85702181d7473ab2a9a4d2d4b869d586c3091f811b12a81a63aae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfzrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d876df04b5ab0ba07a5cbcafbe4648403c5014c59eb7d40adee2657c3a255473\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d876df04b5ab0ba07a5cbcafbe4648403c5014c59eb7d40adee2657c3a255473\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfzrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:23Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-5w4fr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:41Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:41 crc kubenswrapper[4980]: I1206 03:34:41.269624 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:34:41 crc kubenswrapper[4980]: I1206 03:34:41.269674 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:34:41 crc kubenswrapper[4980]: I1206 03:34:41.269690 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:34:41 crc kubenswrapper[4980]: I1206 03:34:41.269713 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:34:41 crc kubenswrapper[4980]: I1206 03:34:41.269730 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:34:41Z","lastTransitionTime":"2025-12-06T03:34:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:34:41 crc kubenswrapper[4980]: I1206 03:34:41.308568 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cab608b3-a02f-475b-9aae-9b8f5d878de9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f7fcc44b42bcbb55509048e3a566bb557333c4a83e33d7b0fba8eb514317909\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c22e7dd8886c2dfbb742900ed86ec7b0c2edf9a1363e74f6666961d928f06c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://473e792b2d0a9644854ee0581c6120fc9c989bc342899ef72b1d056080fd4694\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ee8e936dfc39361d6f99460910591ca8c901a20f0d274bbb6aaebfa8a00ee06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://77bc12768b018625ba11a6bfd66688f926c4b7f3bc1d0e41c6551a18b485ecf3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://27cf07cd96893900da4afecd0d56fd9269bc60c1d23246b4685fcd8756357c19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://27cf07cd96893900da4afecd0d56fd9269bc60c1d23246b4685fcd8756357c19\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c49feb24566a8284378670fa0390710d38c237acc94ffb654f90f3fa4626e4c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c49feb24566a8284378670fa0390710d38c237acc94ffb654f90f3fa4626e4c8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:03Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://bb74d1c3d752a379fa394c450217badb45511fbcafad62f7b618131cab94b29e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bb74d1c3d752a379fa394c450217badb45511fbcafad62f7b618131cab94b29e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:01Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:41Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:41 crc kubenswrapper[4980]: I1206 03:34:41.329341 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://502c727d6baa31f400bee14c34b19abbe38af873f28f6a82a43547431755d981\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:41Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:41 crc kubenswrapper[4980]: I1206 03:34:41.373050 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-dx44q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5bac22a-512c-49d0-90db-47de6aaaefd0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://97469dea2cd5393da0acf784ccc1969c8a4d07da7d51b14cdc955013ee9710ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f7mdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3b5dab476c7f227a6e4a2eae8469432667fb9fad345d9419ba59feab9fd385e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c3b5dab476c7f227a6e4a2eae8469432667fb9fad345d9419ba59feab9fd385e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f7mdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e921499231cbdf26b8918c98f9ad8c322fd7848865caeef048470de3e2c0efba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e921499231cbdf26b8918c98f9ad8c322fd7848865caeef048470de3e2c0efba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f7mdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7cbfe7d8cdb89916c238296592fab45b9e92e2c585baa6d67230decd1c84c5bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7cbfe7d8cdb89916c238296592fab45b9e92e2c585baa6d67230decd1c84c5bb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f7mdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f45cbf23b6b0d208cefc4b493f901bab9d71ccf2f1456507d54fb1f30c4fd48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7f45cbf23b6b0d208cefc4b493f901bab9d71ccf2f1456507d54fb1f30c4fd48\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f7mdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://344ce4b145ff021fcaf163803d8ad8cc6e9565bf7f0992e2653c43db663d8411\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://344ce4b145ff021fcaf163803d8ad8cc6e9565bf7f0992e2653c43db663d8411\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f7mdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://619e56a9937dd6376ab471ba16da30e5d77f443f5bd33610bea5ec85cd0446ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://619e56a9937dd6376ab471ba16da30e5d77f443f5bd33610bea5ec85cd0446ef\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f7mdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:23Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-dx44q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:41Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:41 crc kubenswrapper[4980]: I1206 03:34:41.374162 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:34:41 crc kubenswrapper[4980]: I1206 03:34:41.374216 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:34:41 crc kubenswrapper[4980]: I1206 03:34:41.374236 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:34:41 crc kubenswrapper[4980]: I1206 03:34:41.374264 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:34:41 crc kubenswrapper[4980]: I1206 03:34:41.374281 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:34:41Z","lastTransitionTime":"2025-12-06T03:34:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:34:41 crc kubenswrapper[4980]: I1206 03:34:41.406987 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-r5zfb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"320f44d1-a671-4a91-b328-a8b0fdd8f23a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://92393ecaaef8cfcfd01837671f8cc89803dd570b637b4a373b40b7f487788826\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zhm5t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52e8db644463794041978ced2c0591c13e0ada8e5f1a78a94a0a665218a37b5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zhm5t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:23Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-r5zfb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:41Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:41 crc kubenswrapper[4980]: I1206 03:34:41.447126 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-cvzg8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f1cde706-268d-4d26-818f-2f42bfb37b5f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://86f9a77acae360740be827eb0d1e1516e8a49cb9b197b6f5f9468ac70a101158\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rk86l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:29Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-cvzg8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:41Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:41 crc kubenswrapper[4980]: I1206 03:34:41.477042 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:34:41 crc kubenswrapper[4980]: I1206 03:34:41.477103 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:34:41 crc kubenswrapper[4980]: I1206 03:34:41.477117 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:34:41 crc kubenswrapper[4980]: I1206 03:34:41.477139 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:34:41 crc kubenswrapper[4980]: I1206 03:34:41.477155 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:34:41Z","lastTransitionTime":"2025-12-06T03:34:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:34:41 crc kubenswrapper[4980]: I1206 03:34:41.502040 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cab608b3-a02f-475b-9aae-9b8f5d878de9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f7fcc44b42bcbb55509048e3a566bb557333c4a83e33d7b0fba8eb514317909\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c22e7dd8886c2dfbb742900ed86ec7b0c2edf9a1363e74f6666961d928f06c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://473e792b2d0a9644854ee0581c6120fc9c989bc342899ef72b1d056080fd4694\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ee8e936dfc39361d6f99460910591ca8c901a20f0d274bbb6aaebfa8a00ee06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://77bc12768b018625ba11a6bfd66688f926c4b7f3bc1d0e41c6551a18b485ecf3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://27cf07cd96893900da4afecd0d56fd9269bc60c1d23246b4685fcd8756357c19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://27cf07cd96893900da4afecd0d56fd9269bc60c1d23246b4685fcd8756357c19\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c49feb24566a8284378670fa0390710d38c237acc94ffb654f90f3fa4626e4c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c49feb24566a8284378670fa0390710d38c237acc94ffb654f90f3fa4626e4c8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:03Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://bb74d1c3d752a379fa394c450217badb45511fbcafad62f7b618131cab94b29e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bb74d1c3d752a379fa394c450217badb45511fbcafad62f7b618131cab94b29e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:01Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:41Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:41 crc kubenswrapper[4980]: I1206 03:34:41.529121 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://502c727d6baa31f400bee14c34b19abbe38af873f28f6a82a43547431755d981\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:41Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:41 crc kubenswrapper[4980]: I1206 03:34:41.569832 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-dx44q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5bac22a-512c-49d0-90db-47de6aaaefd0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://97469dea2cd5393da0acf784ccc1969c8a4d07da7d51b14cdc955013ee9710ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f7mdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3b5dab476c7f227a6e4a2eae8469432667fb9fad345d9419ba59feab9fd385e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c3b5dab476c7f227a6e4a2eae8469432667fb9fad345d9419ba59feab9fd385e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f7mdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e921499231cbdf26b8918c98f9ad8c322fd7848865caeef048470de3e2c0efba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e921499231cbdf26b8918c98f9ad8c322fd7848865caeef048470de3e2c0efba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f7mdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7cbfe7d8cdb89916c238296592fab45b9e92e2c585baa6d67230decd1c84c5bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7cbfe7d8cdb89916c238296592fab45b9e92e2c585baa6d67230decd1c84c5bb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f7mdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f45cbf23b6b0d208cefc4b493f901bab9d71ccf2f1456507d54fb1f30c4fd48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7f45cbf23b6b0d208cefc4b493f901bab9d71ccf2f1456507d54fb1f30c4fd48\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f7mdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://344ce4b145ff021fcaf163803d8ad8cc6e9565bf7f0992e2653c43db663d8411\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://344ce4b145ff021fcaf163803d8ad8cc6e9565bf7f0992e2653c43db663d8411\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f7mdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://619e56a9937dd6376ab471ba16da30e5d77f443f5bd33610bea5ec85cd0446ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://619e56a9937dd6376ab471ba16da30e5d77f443f5bd33610bea5ec85cd0446ef\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f7mdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:23Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-dx44q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:41Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:41 crc kubenswrapper[4980]: I1206 03:34:41.580832 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:34:41 crc kubenswrapper[4980]: I1206 03:34:41.580890 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:34:41 crc kubenswrapper[4980]: I1206 03:34:41.580912 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:34:41 crc kubenswrapper[4980]: I1206 03:34:41.580974 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:34:41 crc kubenswrapper[4980]: I1206 03:34:41.581000 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:34:41Z","lastTransitionTime":"2025-12-06T03:34:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:34:41 crc kubenswrapper[4980]: I1206 03:34:41.609215 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-r5zfb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"320f44d1-a671-4a91-b328-a8b0fdd8f23a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://92393ecaaef8cfcfd01837671f8cc89803dd570b637b4a373b40b7f487788826\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zhm5t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52e8db644463794041978ced2c0591c13e0ada8e5f1a78a94a0a665218a37b5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zhm5t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:23Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-r5zfb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:41Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:41 crc kubenswrapper[4980]: I1206 03:34:41.645208 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-cvzg8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f1cde706-268d-4d26-818f-2f42bfb37b5f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://86f9a77acae360740be827eb0d1e1516e8a49cb9b197b6f5f9468ac70a101158\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rk86l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:29Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-cvzg8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:41Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:41 crc kubenswrapper[4980]: I1206 03:34:41.684178 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:34:41 crc kubenswrapper[4980]: I1206 03:34:41.684222 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:34:41 crc kubenswrapper[4980]: I1206 03:34:41.684233 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:34:41 crc kubenswrapper[4980]: I1206 03:34:41.684253 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:34:41 crc kubenswrapper[4980]: I1206 03:34:41.684267 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:34:41Z","lastTransitionTime":"2025-12-06T03:34:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:34:41 crc kubenswrapper[4980]: I1206 03:34:41.689589 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:41Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:41 crc kubenswrapper[4980]: I1206 03:34:41.731327 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4ef01e7d9809fd9cfc46e3a58b07159cdc6f5a3c210d1c9efc49bca27b30858\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:41Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:41 crc kubenswrapper[4980]: I1206 03:34:41.768821 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-kq4z8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f96c1f6c-55e8-4d86-bff3-5381581b7022\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://799498681afa7653408bc23b1e82d75c16e3c809d2ee7f322780e20c537eb710\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bdx7g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:23Z\\\"}}\" for pod \"openshift-multus\"/\"multus-kq4z8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:41Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:41 crc kubenswrapper[4980]: I1206 03:34:41.786559 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:34:41 crc kubenswrapper[4980]: I1206 03:34:41.786631 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:34:41 crc kubenswrapper[4980]: I1206 03:34:41.786655 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:34:41 crc kubenswrapper[4980]: I1206 03:34:41.786684 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:34:41 crc kubenswrapper[4980]: I1206 03:34:41.786702 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:34:41Z","lastTransitionTime":"2025-12-06T03:34:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:34:41 crc kubenswrapper[4980]: I1206 03:34:41.807434 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-6kxfq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"951b6deb-bcea-4332-8ae8-c23ac7aaef36\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://423d30c3e65b04bdf6ff5ef53d8a7575a434c05e28489370a7370381ebc29de9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b8htf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://282df50dd9653c1cbe97fc2041d9e9ea91a992661682652aafc12e566d8f1b81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b8htf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:37Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-6kxfq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:41Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:41 crc kubenswrapper[4980]: I1206 03:34:41.847452 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-mcprh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e6e88c5e-a255-449a-ae38-ac4d73a8e984\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g7xz9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g7xz9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:38Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-mcprh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:41Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:41 crc kubenswrapper[4980]: I1206 03:34:41.890749 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:34:41 crc kubenswrapper[4980]: I1206 03:34:41.890847 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:34:41 crc kubenswrapper[4980]: I1206 03:34:41.890876 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:34:41 crc kubenswrapper[4980]: I1206 03:34:41.890911 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:34:41 crc kubenswrapper[4980]: I1206 03:34:41.890937 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:34:41Z","lastTransitionTime":"2025-12-06T03:34:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:34:41 crc kubenswrapper[4980]: I1206 03:34:41.892835 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7f958150-549d-4c80-8a7e-23c39e4f9500\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c28b6895b14e73f4f3eb7a8ce93076c05092ba9b58ea0e3e879c65bcfa227a21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://76cff8a2d652df38b04eefc053832baf4cbb5db778ff5e525b18c2825141b118\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://206f6f19fe3197f5d6798cc86c5473bf2decb02a2860ee7ffb5efb5f60952e8c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bafaa9670b05671fb57b658c2960bf3ee32c84e52f86a3a3bbab01d5a38ff0c4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://773a0bc2aab73ffc2717e987543defc9b786cc22300aa54b4ba4f3454504e331\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-06T03:34:22Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1206 03:34:15.305330 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1206 03:34:15.307249 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1592846546/tls.crt::/tmp/serving-cert-1592846546/tls.key\\\\\\\"\\\\nI1206 03:34:21.814037 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1206 03:34:21.816756 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1206 03:34:21.816780 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1206 03:34:21.816806 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1206 03:34:21.816813 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1206 03:34:21.831301 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1206 03:34:21.831311 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1206 03:34:21.831324 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 03:34:21.831330 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 03:34:21.831335 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1206 03:34:21.831339 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1206 03:34:21.831342 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1206 03:34:21.831346 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1206 03:34:21.911269 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:04Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d40e0e7446f8416bf9da94fc53c01be937f9a51a11eb101157d2ddd4597720d2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:04Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2cba89f2f1bed96d59f183628d900d9b72c91eb472e273c8bed0babcab485910\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2cba89f2f1bed96d59f183628d900d9b72c91eb472e273c8bed0babcab485910\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:01Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:41Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:41 crc kubenswrapper[4980]: I1206 03:34:41.935146 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6ed68f66-1cf6-465a-8f66-76d5af6a44d8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://725e84f6bdfb722a8ce69158d03ea4c66a9f087642d55cb9651e003c81be0c8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://71f7b492ea07cdf7dc65fc48184ec1e64914eb1811bbfdddc0fb14c8ad6cb296\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b543aa4ee03d5a14909b67d876c5675c02e1df2a89a3c9ac8d3fb587fea53d8f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://32af3427e872a70cb3c2357116f04cabc8d3d9b69e7306b2e0de2ae22f9469d4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:01Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:41Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:41 crc kubenswrapper[4980]: I1206 03:34:41.983577 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5w4fr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5af0e768-3c9c-4401-ab56-57516cd3170e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:23Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:23Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://768b8e1fe1a00a8467d744da5ad3f61006b0a3612c96f8e4d0ee6fcb93195f9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfzrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6342903b318317ee8d84c2f5d467cb25c811a789be0d6ab0e0cba37b5611e16d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfzrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d620685574db7594132ae24a91ec0a93dce20a9a289540deebb39650b8a1816\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfzrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c67ec8f9c1764fcdf215d52822c2d4f0437dc7f056f05d708767121f6c1f3515\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfzrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f35beae5339eeaf26494ed678e3c1bd02349eae2a4b4615325b17923f3b653b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfzrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a940ec6ee523313728f7e8ae934b7e0b0cff82771e3e64723309db1fae013c3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfzrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6cb5cc9f6f22133cebf095db595e5062d799d7e3c8ca0bf31d3809a296e507cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6cb5cc9f6f22133cebf095db595e5062d799d7e3c8ca0bf31d3809a296e507cc\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-06T03:34:39Z\\\",\\\"message\\\":\\\"lector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1206 03:34:39.306579 6370 reflector.go:311] Stopping reflector *v1.AdminPolicyBasedExternalRoute (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/adminpolicybasedroute/v1/apis/informers/externalversions/factory.go:140\\\\nI1206 03:34:39.306665 6370 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1206 03:34:39.306723 6370 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI1206 03:34:39.306893 6370 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1206 03:34:39.307227 6370 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1206 03:34:39.307306 6370 reflector.go:311] Stopping reflector *v1.NetworkAttachmentDefinition (0s) from github.com/k8snetworkplumbingwg/network-attachment-definition-client/pkg/client/informers/externalversions/factory.go:117\\\\nI1206 03:34:39.307466 6370 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:37Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-5w4fr_openshift-ovn-kubernetes(5af0e768-3c9c-4401-ab56-57516cd3170e)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfzrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe261f4322b85702181d7473ab2a9a4d2d4b869d586c3091f811b12a81a63aae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfzrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d876df04b5ab0ba07a5cbcafbe4648403c5014c59eb7d40adee2657c3a255473\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d876df04b5ab0ba07a5cbcafbe4648403c5014c59eb7d40adee2657c3a255473\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfzrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:23Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-5w4fr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:41Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:41 crc kubenswrapper[4980]: I1206 03:34:41.995477 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:34:41 crc kubenswrapper[4980]: I1206 03:34:41.995594 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:34:41 crc kubenswrapper[4980]: I1206 03:34:41.995651 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:34:41 crc kubenswrapper[4980]: I1206 03:34:41.995675 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:34:41 crc kubenswrapper[4980]: I1206 03:34:41.995690 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:34:41Z","lastTransitionTime":"2025-12-06T03:34:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:34:42 crc kubenswrapper[4980]: I1206 03:34:42.007570 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:22Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:42Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:42 crc kubenswrapper[4980]: I1206 03:34:42.050972 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://334ea6a273fd7d268906c9f4a6dbf6f4df4fcdbf38f3d7d248cb7d6cf8537794\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ec44dcd5e4bf0a0466bb586ebd61259364dbbfe22499f4cbff854e3bd153082\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:42Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:42 crc kubenswrapper[4980]: I1206 03:34:42.092954 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:22Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:42Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:42 crc kubenswrapper[4980]: I1206 03:34:42.099080 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:34:42 crc kubenswrapper[4980]: I1206 03:34:42.099163 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:34:42 crc kubenswrapper[4980]: I1206 03:34:42.099191 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:34:42 crc kubenswrapper[4980]: I1206 03:34:42.099226 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:34:42 crc kubenswrapper[4980]: I1206 03:34:42.099249 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:34:42Z","lastTransitionTime":"2025-12-06T03:34:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:34:42 crc kubenswrapper[4980]: I1206 03:34:42.128850 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-sv5sz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"873977aa-4eb0-4526-bc0b-757ad2dea2a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01f0cc855c822a71ea8ef6a7ce6e25747ff9956691735973f935c4a28672a7a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lg7tx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:23Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-sv5sz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:42Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:42 crc kubenswrapper[4980]: I1206 03:34:42.183999 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 03:34:42 crc kubenswrapper[4980]: I1206 03:34:42.184069 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 03:34:42 crc kubenswrapper[4980]: E1206 03:34:42.184142 4980 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 03:34:42 crc kubenswrapper[4980]: I1206 03:34:42.183999 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-mcprh" Dec 06 03:34:42 crc kubenswrapper[4980]: E1206 03:34:42.184326 4980 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 03:34:42 crc kubenswrapper[4980]: E1206 03:34:42.184480 4980 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-mcprh" podUID="e6e88c5e-a255-449a-ae38-ac4d73a8e984" Dec 06 03:34:42 crc kubenswrapper[4980]: I1206 03:34:42.202378 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:34:42 crc kubenswrapper[4980]: I1206 03:34:42.202462 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:34:42 crc kubenswrapper[4980]: I1206 03:34:42.202488 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:34:42 crc kubenswrapper[4980]: I1206 03:34:42.202620 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:34:42 crc kubenswrapper[4980]: I1206 03:34:42.202647 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:34:42Z","lastTransitionTime":"2025-12-06T03:34:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:34:42 crc kubenswrapper[4980]: I1206 03:34:42.306810 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:34:42 crc kubenswrapper[4980]: I1206 03:34:42.306906 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:34:42 crc kubenswrapper[4980]: I1206 03:34:42.306916 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:34:42 crc kubenswrapper[4980]: I1206 03:34:42.306946 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:34:42 crc kubenswrapper[4980]: I1206 03:34:42.306964 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:34:42Z","lastTransitionTime":"2025-12-06T03:34:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:34:42 crc kubenswrapper[4980]: I1206 03:34:42.337895 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/e6e88c5e-a255-449a-ae38-ac4d73a8e984-metrics-certs\") pod \"network-metrics-daemon-mcprh\" (UID: \"e6e88c5e-a255-449a-ae38-ac4d73a8e984\") " pod="openshift-multus/network-metrics-daemon-mcprh" Dec 06 03:34:42 crc kubenswrapper[4980]: E1206 03:34:42.338123 4980 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 06 03:34:42 crc kubenswrapper[4980]: E1206 03:34:42.338225 4980 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/e6e88c5e-a255-449a-ae38-ac4d73a8e984-metrics-certs podName:e6e88c5e-a255-449a-ae38-ac4d73a8e984 nodeName:}" failed. No retries permitted until 2025-12-06 03:34:46.338200602 +0000 UTC m=+45.582951873 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/e6e88c5e-a255-449a-ae38-ac4d73a8e984-metrics-certs") pod "network-metrics-daemon-mcprh" (UID: "e6e88c5e-a255-449a-ae38-ac4d73a8e984") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 06 03:34:42 crc kubenswrapper[4980]: I1206 03:34:42.411460 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:34:42 crc kubenswrapper[4980]: I1206 03:34:42.411595 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:34:42 crc kubenswrapper[4980]: I1206 03:34:42.411631 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:34:42 crc kubenswrapper[4980]: I1206 03:34:42.411752 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:34:42 crc kubenswrapper[4980]: I1206 03:34:42.411830 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:34:42Z","lastTransitionTime":"2025-12-06T03:34:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:34:42 crc kubenswrapper[4980]: I1206 03:34:42.515807 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:34:42 crc kubenswrapper[4980]: I1206 03:34:42.515880 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:34:42 crc kubenswrapper[4980]: I1206 03:34:42.515898 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:34:42 crc kubenswrapper[4980]: I1206 03:34:42.515926 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:34:42 crc kubenswrapper[4980]: I1206 03:34:42.515945 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:34:42Z","lastTransitionTime":"2025-12-06T03:34:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:34:42 crc kubenswrapper[4980]: I1206 03:34:42.619389 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:34:42 crc kubenswrapper[4980]: I1206 03:34:42.619462 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:34:42 crc kubenswrapper[4980]: I1206 03:34:42.619479 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:34:42 crc kubenswrapper[4980]: I1206 03:34:42.619506 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:34:42 crc kubenswrapper[4980]: I1206 03:34:42.619552 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:34:42Z","lastTransitionTime":"2025-12-06T03:34:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:34:42 crc kubenswrapper[4980]: I1206 03:34:42.724033 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:34:42 crc kubenswrapper[4980]: I1206 03:34:42.724109 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:34:42 crc kubenswrapper[4980]: I1206 03:34:42.724127 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:34:42 crc kubenswrapper[4980]: I1206 03:34:42.724154 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:34:42 crc kubenswrapper[4980]: I1206 03:34:42.724178 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:34:42Z","lastTransitionTime":"2025-12-06T03:34:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:34:42 crc kubenswrapper[4980]: I1206 03:34:42.827717 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:34:42 crc kubenswrapper[4980]: I1206 03:34:42.827786 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:34:42 crc kubenswrapper[4980]: I1206 03:34:42.827803 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:34:42 crc kubenswrapper[4980]: I1206 03:34:42.827828 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:34:42 crc kubenswrapper[4980]: I1206 03:34:42.827846 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:34:42Z","lastTransitionTime":"2025-12-06T03:34:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:34:42 crc kubenswrapper[4980]: I1206 03:34:42.930830 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:34:42 crc kubenswrapper[4980]: I1206 03:34:42.931070 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:34:42 crc kubenswrapper[4980]: I1206 03:34:42.931090 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:34:42 crc kubenswrapper[4980]: I1206 03:34:42.931114 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:34:42 crc kubenswrapper[4980]: I1206 03:34:42.931131 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:34:42Z","lastTransitionTime":"2025-12-06T03:34:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:34:43 crc kubenswrapper[4980]: I1206 03:34:43.034493 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:34:43 crc kubenswrapper[4980]: I1206 03:34:43.034600 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:34:43 crc kubenswrapper[4980]: I1206 03:34:43.034631 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:34:43 crc kubenswrapper[4980]: I1206 03:34:43.034662 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:34:43 crc kubenswrapper[4980]: I1206 03:34:43.034685 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:34:43Z","lastTransitionTime":"2025-12-06T03:34:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:34:43 crc kubenswrapper[4980]: I1206 03:34:43.138150 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:34:43 crc kubenswrapper[4980]: I1206 03:34:43.138199 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:34:43 crc kubenswrapper[4980]: I1206 03:34:43.138215 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:34:43 crc kubenswrapper[4980]: I1206 03:34:43.138237 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:34:43 crc kubenswrapper[4980]: I1206 03:34:43.138256 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:34:43Z","lastTransitionTime":"2025-12-06T03:34:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:34:43 crc kubenswrapper[4980]: I1206 03:34:43.183666 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 03:34:43 crc kubenswrapper[4980]: E1206 03:34:43.183837 4980 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 03:34:43 crc kubenswrapper[4980]: I1206 03:34:43.241864 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:34:43 crc kubenswrapper[4980]: I1206 03:34:43.241991 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:34:43 crc kubenswrapper[4980]: I1206 03:34:43.242022 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:34:43 crc kubenswrapper[4980]: I1206 03:34:43.242056 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:34:43 crc kubenswrapper[4980]: I1206 03:34:43.242081 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:34:43Z","lastTransitionTime":"2025-12-06T03:34:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:34:43 crc kubenswrapper[4980]: I1206 03:34:43.345739 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:34:43 crc kubenswrapper[4980]: I1206 03:34:43.345826 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:34:43 crc kubenswrapper[4980]: I1206 03:34:43.345847 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:34:43 crc kubenswrapper[4980]: I1206 03:34:43.345873 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:34:43 crc kubenswrapper[4980]: I1206 03:34:43.345899 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:34:43Z","lastTransitionTime":"2025-12-06T03:34:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:34:43 crc kubenswrapper[4980]: I1206 03:34:43.449732 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:34:43 crc kubenswrapper[4980]: I1206 03:34:43.449804 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:34:43 crc kubenswrapper[4980]: I1206 03:34:43.449826 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:34:43 crc kubenswrapper[4980]: I1206 03:34:43.449927 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:34:43 crc kubenswrapper[4980]: I1206 03:34:43.449957 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:34:43Z","lastTransitionTime":"2025-12-06T03:34:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:34:43 crc kubenswrapper[4980]: I1206 03:34:43.553772 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:34:43 crc kubenswrapper[4980]: I1206 03:34:43.553888 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:34:43 crc kubenswrapper[4980]: I1206 03:34:43.553901 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:34:43 crc kubenswrapper[4980]: I1206 03:34:43.553918 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:34:43 crc kubenswrapper[4980]: I1206 03:34:43.553928 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:34:43Z","lastTransitionTime":"2025-12-06T03:34:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:34:43 crc kubenswrapper[4980]: I1206 03:34:43.658137 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:34:43 crc kubenswrapper[4980]: I1206 03:34:43.658260 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:34:43 crc kubenswrapper[4980]: I1206 03:34:43.658330 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:34:43 crc kubenswrapper[4980]: I1206 03:34:43.658364 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:34:43 crc kubenswrapper[4980]: I1206 03:34:43.658431 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:34:43Z","lastTransitionTime":"2025-12-06T03:34:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:34:43 crc kubenswrapper[4980]: I1206 03:34:43.762462 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:34:43 crc kubenswrapper[4980]: I1206 03:34:43.762554 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:34:43 crc kubenswrapper[4980]: I1206 03:34:43.762574 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:34:43 crc kubenswrapper[4980]: I1206 03:34:43.762598 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:34:43 crc kubenswrapper[4980]: I1206 03:34:43.762617 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:34:43Z","lastTransitionTime":"2025-12-06T03:34:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:34:43 crc kubenswrapper[4980]: I1206 03:34:43.866753 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:34:43 crc kubenswrapper[4980]: I1206 03:34:43.866827 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:34:43 crc kubenswrapper[4980]: I1206 03:34:43.866845 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:34:43 crc kubenswrapper[4980]: I1206 03:34:43.866870 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:34:43 crc kubenswrapper[4980]: I1206 03:34:43.866887 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:34:43Z","lastTransitionTime":"2025-12-06T03:34:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:34:43 crc kubenswrapper[4980]: I1206 03:34:43.970339 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:34:43 crc kubenswrapper[4980]: I1206 03:34:43.970455 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:34:43 crc kubenswrapper[4980]: I1206 03:34:43.970471 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:34:43 crc kubenswrapper[4980]: I1206 03:34:43.970490 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:34:43 crc kubenswrapper[4980]: I1206 03:34:43.970535 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:34:43Z","lastTransitionTime":"2025-12-06T03:34:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:34:44 crc kubenswrapper[4980]: I1206 03:34:44.073456 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:34:44 crc kubenswrapper[4980]: I1206 03:34:44.073560 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:34:44 crc kubenswrapper[4980]: I1206 03:34:44.073599 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:34:44 crc kubenswrapper[4980]: I1206 03:34:44.073633 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:34:44 crc kubenswrapper[4980]: I1206 03:34:44.073655 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:34:44Z","lastTransitionTime":"2025-12-06T03:34:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:34:44 crc kubenswrapper[4980]: I1206 03:34:44.176997 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:34:44 crc kubenswrapper[4980]: I1206 03:34:44.177055 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:34:44 crc kubenswrapper[4980]: I1206 03:34:44.177071 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:34:44 crc kubenswrapper[4980]: I1206 03:34:44.177094 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:34:44 crc kubenswrapper[4980]: I1206 03:34:44.177112 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:34:44Z","lastTransitionTime":"2025-12-06T03:34:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:34:44 crc kubenswrapper[4980]: I1206 03:34:44.184188 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 03:34:44 crc kubenswrapper[4980]: I1206 03:34:44.184257 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 03:34:44 crc kubenswrapper[4980]: E1206 03:34:44.184431 4980 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 03:34:44 crc kubenswrapper[4980]: I1206 03:34:44.184268 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-mcprh" Dec 06 03:34:44 crc kubenswrapper[4980]: E1206 03:34:44.184579 4980 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 03:34:44 crc kubenswrapper[4980]: E1206 03:34:44.184615 4980 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-mcprh" podUID="e6e88c5e-a255-449a-ae38-ac4d73a8e984" Dec 06 03:34:44 crc kubenswrapper[4980]: I1206 03:34:44.281602 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:34:44 crc kubenswrapper[4980]: I1206 03:34:44.281663 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:34:44 crc kubenswrapper[4980]: I1206 03:34:44.281682 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:34:44 crc kubenswrapper[4980]: I1206 03:34:44.281725 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:34:44 crc kubenswrapper[4980]: I1206 03:34:44.281741 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:34:44Z","lastTransitionTime":"2025-12-06T03:34:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:34:44 crc kubenswrapper[4980]: I1206 03:34:44.385924 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:34:44 crc kubenswrapper[4980]: I1206 03:34:44.386012 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:34:44 crc kubenswrapper[4980]: I1206 03:34:44.386030 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:34:44 crc kubenswrapper[4980]: I1206 03:34:44.386057 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:34:44 crc kubenswrapper[4980]: I1206 03:34:44.386075 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:34:44Z","lastTransitionTime":"2025-12-06T03:34:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:34:44 crc kubenswrapper[4980]: I1206 03:34:44.489799 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:34:44 crc kubenswrapper[4980]: I1206 03:34:44.489872 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:34:44 crc kubenswrapper[4980]: I1206 03:34:44.489891 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:34:44 crc kubenswrapper[4980]: I1206 03:34:44.489920 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:34:44 crc kubenswrapper[4980]: I1206 03:34:44.489941 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:34:44Z","lastTransitionTime":"2025-12-06T03:34:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:34:44 crc kubenswrapper[4980]: I1206 03:34:44.592900 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:34:44 crc kubenswrapper[4980]: I1206 03:34:44.592982 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:34:44 crc kubenswrapper[4980]: I1206 03:34:44.593002 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:34:44 crc kubenswrapper[4980]: I1206 03:34:44.593034 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:34:44 crc kubenswrapper[4980]: I1206 03:34:44.593056 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:34:44Z","lastTransitionTime":"2025-12-06T03:34:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:34:44 crc kubenswrapper[4980]: I1206 03:34:44.696252 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:34:44 crc kubenswrapper[4980]: I1206 03:34:44.696324 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:34:44 crc kubenswrapper[4980]: I1206 03:34:44.696384 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:34:44 crc kubenswrapper[4980]: I1206 03:34:44.696414 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:34:44 crc kubenswrapper[4980]: I1206 03:34:44.696434 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:34:44Z","lastTransitionTime":"2025-12-06T03:34:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:34:44 crc kubenswrapper[4980]: I1206 03:34:44.799446 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:34:44 crc kubenswrapper[4980]: I1206 03:34:44.799535 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:34:44 crc kubenswrapper[4980]: I1206 03:34:44.799558 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:34:44 crc kubenswrapper[4980]: I1206 03:34:44.799584 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:34:44 crc kubenswrapper[4980]: I1206 03:34:44.799604 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:34:44Z","lastTransitionTime":"2025-12-06T03:34:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:34:44 crc kubenswrapper[4980]: I1206 03:34:44.902104 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:34:44 crc kubenswrapper[4980]: I1206 03:34:44.902147 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:34:44 crc kubenswrapper[4980]: I1206 03:34:44.902158 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:34:44 crc kubenswrapper[4980]: I1206 03:34:44.902174 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:34:44 crc kubenswrapper[4980]: I1206 03:34:44.902185 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:34:44Z","lastTransitionTime":"2025-12-06T03:34:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:34:45 crc kubenswrapper[4980]: I1206 03:34:45.005707 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:34:45 crc kubenswrapper[4980]: I1206 03:34:45.005767 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:34:45 crc kubenswrapper[4980]: I1206 03:34:45.005784 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:34:45 crc kubenswrapper[4980]: I1206 03:34:45.005807 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:34:45 crc kubenswrapper[4980]: I1206 03:34:45.005822 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:34:45Z","lastTransitionTime":"2025-12-06T03:34:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:34:45 crc kubenswrapper[4980]: I1206 03:34:45.109086 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:34:45 crc kubenswrapper[4980]: I1206 03:34:45.109174 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:34:45 crc kubenswrapper[4980]: I1206 03:34:45.109201 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:34:45 crc kubenswrapper[4980]: I1206 03:34:45.109234 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:34:45 crc kubenswrapper[4980]: I1206 03:34:45.109256 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:34:45Z","lastTransitionTime":"2025-12-06T03:34:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:34:45 crc kubenswrapper[4980]: I1206 03:34:45.184105 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 03:34:45 crc kubenswrapper[4980]: E1206 03:34:45.184297 4980 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 03:34:45 crc kubenswrapper[4980]: I1206 03:34:45.212134 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:34:45 crc kubenswrapper[4980]: I1206 03:34:45.212212 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:34:45 crc kubenswrapper[4980]: I1206 03:34:45.212234 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:34:45 crc kubenswrapper[4980]: I1206 03:34:45.212260 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:34:45 crc kubenswrapper[4980]: I1206 03:34:45.212281 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:34:45Z","lastTransitionTime":"2025-12-06T03:34:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:34:45 crc kubenswrapper[4980]: I1206 03:34:45.315301 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:34:45 crc kubenswrapper[4980]: I1206 03:34:45.315371 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:34:45 crc kubenswrapper[4980]: I1206 03:34:45.315409 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:34:45 crc kubenswrapper[4980]: I1206 03:34:45.315441 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:34:45 crc kubenswrapper[4980]: I1206 03:34:45.315501 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:34:45Z","lastTransitionTime":"2025-12-06T03:34:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:34:45 crc kubenswrapper[4980]: I1206 03:34:45.420086 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:34:45 crc kubenswrapper[4980]: I1206 03:34:45.420165 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:34:45 crc kubenswrapper[4980]: I1206 03:34:45.420180 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:34:45 crc kubenswrapper[4980]: I1206 03:34:45.420206 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:34:45 crc kubenswrapper[4980]: I1206 03:34:45.420228 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:34:45Z","lastTransitionTime":"2025-12-06T03:34:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:34:45 crc kubenswrapper[4980]: I1206 03:34:45.524460 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:34:45 crc kubenswrapper[4980]: I1206 03:34:45.524548 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:34:45 crc kubenswrapper[4980]: I1206 03:34:45.524568 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:34:45 crc kubenswrapper[4980]: I1206 03:34:45.524597 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:34:45 crc kubenswrapper[4980]: I1206 03:34:45.524615 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:34:45Z","lastTransitionTime":"2025-12-06T03:34:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:34:45 crc kubenswrapper[4980]: I1206 03:34:45.628022 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:34:45 crc kubenswrapper[4980]: I1206 03:34:45.628100 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:34:45 crc kubenswrapper[4980]: I1206 03:34:45.628124 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:34:45 crc kubenswrapper[4980]: I1206 03:34:45.628170 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:34:45 crc kubenswrapper[4980]: I1206 03:34:45.628191 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:34:45Z","lastTransitionTime":"2025-12-06T03:34:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:34:45 crc kubenswrapper[4980]: I1206 03:34:45.732651 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:34:45 crc kubenswrapper[4980]: I1206 03:34:45.732725 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:34:45 crc kubenswrapper[4980]: I1206 03:34:45.732736 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:34:45 crc kubenswrapper[4980]: I1206 03:34:45.732760 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:34:45 crc kubenswrapper[4980]: I1206 03:34:45.732774 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:34:45Z","lastTransitionTime":"2025-12-06T03:34:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:34:45 crc kubenswrapper[4980]: I1206 03:34:45.836164 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:34:45 crc kubenswrapper[4980]: I1206 03:34:45.836216 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:34:45 crc kubenswrapper[4980]: I1206 03:34:45.836234 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:34:45 crc kubenswrapper[4980]: I1206 03:34:45.836266 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:34:45 crc kubenswrapper[4980]: I1206 03:34:45.836282 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:34:45Z","lastTransitionTime":"2025-12-06T03:34:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:34:45 crc kubenswrapper[4980]: I1206 03:34:45.940329 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:34:45 crc kubenswrapper[4980]: I1206 03:34:45.940398 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:34:45 crc kubenswrapper[4980]: I1206 03:34:45.940417 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:34:45 crc kubenswrapper[4980]: I1206 03:34:45.940444 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:34:45 crc kubenswrapper[4980]: I1206 03:34:45.940461 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:34:45Z","lastTransitionTime":"2025-12-06T03:34:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:34:46 crc kubenswrapper[4980]: I1206 03:34:46.043318 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:34:46 crc kubenswrapper[4980]: I1206 03:34:46.043415 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:34:46 crc kubenswrapper[4980]: I1206 03:34:46.043434 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:34:46 crc kubenswrapper[4980]: I1206 03:34:46.043461 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:34:46 crc kubenswrapper[4980]: I1206 03:34:46.043481 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:34:46Z","lastTransitionTime":"2025-12-06T03:34:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:34:46 crc kubenswrapper[4980]: I1206 03:34:46.147037 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:34:46 crc kubenswrapper[4980]: I1206 03:34:46.147118 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:34:46 crc kubenswrapper[4980]: I1206 03:34:46.147129 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:34:46 crc kubenswrapper[4980]: I1206 03:34:46.147171 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:34:46 crc kubenswrapper[4980]: I1206 03:34:46.147189 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:34:46Z","lastTransitionTime":"2025-12-06T03:34:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:34:46 crc kubenswrapper[4980]: I1206 03:34:46.463187 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/e6e88c5e-a255-449a-ae38-ac4d73a8e984-metrics-certs\") pod \"network-metrics-daemon-mcprh\" (UID: \"e6e88c5e-a255-449a-ae38-ac4d73a8e984\") " pod="openshift-multus/network-metrics-daemon-mcprh" Dec 06 03:34:46 crc kubenswrapper[4980]: E1206 03:34:46.464082 4980 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 06 03:34:46 crc kubenswrapper[4980]: E1206 03:34:46.464271 4980 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/e6e88c5e-a255-449a-ae38-ac4d73a8e984-metrics-certs podName:e6e88c5e-a255-449a-ae38-ac4d73a8e984 nodeName:}" failed. No retries permitted until 2025-12-06 03:34:54.464234555 +0000 UTC m=+53.708985866 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/e6e88c5e-a255-449a-ae38-ac4d73a8e984-metrics-certs") pod "network-metrics-daemon-mcprh" (UID: "e6e88c5e-a255-449a-ae38-ac4d73a8e984") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 06 03:34:46 crc kubenswrapper[4980]: I1206 03:34:46.470672 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 03:34:46 crc kubenswrapper[4980]: E1206 03:34:46.470903 4980 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 03:34:46 crc kubenswrapper[4980]: I1206 03:34:46.470502 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-mcprh" Dec 06 03:34:46 crc kubenswrapper[4980]: E1206 03:34:46.471674 4980 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-mcprh" podUID="e6e88c5e-a255-449a-ae38-ac4d73a8e984" Dec 06 03:34:46 crc kubenswrapper[4980]: I1206 03:34:46.471785 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 03:34:46 crc kubenswrapper[4980]: E1206 03:34:46.471875 4980 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 03:34:46 crc kubenswrapper[4980]: I1206 03:34:46.472940 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:34:46 crc kubenswrapper[4980]: I1206 03:34:46.472990 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:34:46 crc kubenswrapper[4980]: I1206 03:34:46.473011 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:34:46 crc kubenswrapper[4980]: I1206 03:34:46.473036 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:34:46 crc kubenswrapper[4980]: I1206 03:34:46.473053 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:34:46Z","lastTransitionTime":"2025-12-06T03:34:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:34:46 crc kubenswrapper[4980]: I1206 03:34:46.576016 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:34:46 crc kubenswrapper[4980]: I1206 03:34:46.576066 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:34:46 crc kubenswrapper[4980]: I1206 03:34:46.576081 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:34:46 crc kubenswrapper[4980]: I1206 03:34:46.576101 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:34:46 crc kubenswrapper[4980]: I1206 03:34:46.576115 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:34:46Z","lastTransitionTime":"2025-12-06T03:34:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:34:46 crc kubenswrapper[4980]: I1206 03:34:46.679031 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:34:46 crc kubenswrapper[4980]: I1206 03:34:46.679425 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:34:46 crc kubenswrapper[4980]: I1206 03:34:46.679529 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:34:46 crc kubenswrapper[4980]: I1206 03:34:46.679619 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:34:46 crc kubenswrapper[4980]: I1206 03:34:46.679758 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:34:46Z","lastTransitionTime":"2025-12-06T03:34:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:34:46 crc kubenswrapper[4980]: I1206 03:34:46.783010 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:34:46 crc kubenswrapper[4980]: I1206 03:34:46.783463 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:34:46 crc kubenswrapper[4980]: I1206 03:34:46.783557 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:34:46 crc kubenswrapper[4980]: I1206 03:34:46.783641 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:34:46 crc kubenswrapper[4980]: I1206 03:34:46.783719 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:34:46Z","lastTransitionTime":"2025-12-06T03:34:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:34:46 crc kubenswrapper[4980]: I1206 03:34:46.886962 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:34:46 crc kubenswrapper[4980]: I1206 03:34:46.887023 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:34:46 crc kubenswrapper[4980]: I1206 03:34:46.887035 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:34:46 crc kubenswrapper[4980]: I1206 03:34:46.887056 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:34:46 crc kubenswrapper[4980]: I1206 03:34:46.887071 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:34:46Z","lastTransitionTime":"2025-12-06T03:34:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:34:46 crc kubenswrapper[4980]: I1206 03:34:46.990463 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:34:46 crc kubenswrapper[4980]: I1206 03:34:46.991301 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:34:46 crc kubenswrapper[4980]: I1206 03:34:46.991403 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:34:46 crc kubenswrapper[4980]: I1206 03:34:46.991532 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:34:46 crc kubenswrapper[4980]: I1206 03:34:46.991629 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:34:46Z","lastTransitionTime":"2025-12-06T03:34:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:34:47 crc kubenswrapper[4980]: I1206 03:34:47.094945 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:34:47 crc kubenswrapper[4980]: I1206 03:34:47.095019 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:34:47 crc kubenswrapper[4980]: I1206 03:34:47.095033 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:34:47 crc kubenswrapper[4980]: I1206 03:34:47.095054 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:34:47 crc kubenswrapper[4980]: I1206 03:34:47.095068 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:34:47Z","lastTransitionTime":"2025-12-06T03:34:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:34:47 crc kubenswrapper[4980]: I1206 03:34:47.183907 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 03:34:47 crc kubenswrapper[4980]: E1206 03:34:47.184082 4980 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 03:34:47 crc kubenswrapper[4980]: I1206 03:34:47.197621 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:34:47 crc kubenswrapper[4980]: I1206 03:34:47.197671 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:34:47 crc kubenswrapper[4980]: I1206 03:34:47.197683 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:34:47 crc kubenswrapper[4980]: I1206 03:34:47.197702 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:34:47 crc kubenswrapper[4980]: I1206 03:34:47.197718 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:34:47Z","lastTransitionTime":"2025-12-06T03:34:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:34:47 crc kubenswrapper[4980]: I1206 03:34:47.301246 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:34:47 crc kubenswrapper[4980]: I1206 03:34:47.301761 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:34:47 crc kubenswrapper[4980]: I1206 03:34:47.301855 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:34:47 crc kubenswrapper[4980]: I1206 03:34:47.301978 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:34:47 crc kubenswrapper[4980]: I1206 03:34:47.302069 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:34:47Z","lastTransitionTime":"2025-12-06T03:34:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:34:47 crc kubenswrapper[4980]: I1206 03:34:47.405533 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:34:47 crc kubenswrapper[4980]: I1206 03:34:47.405596 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:34:47 crc kubenswrapper[4980]: I1206 03:34:47.405615 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:34:47 crc kubenswrapper[4980]: I1206 03:34:47.405641 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:34:47 crc kubenswrapper[4980]: I1206 03:34:47.405663 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:34:47Z","lastTransitionTime":"2025-12-06T03:34:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:34:47 crc kubenswrapper[4980]: I1206 03:34:47.509012 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:34:47 crc kubenswrapper[4980]: I1206 03:34:47.509080 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:34:47 crc kubenswrapper[4980]: I1206 03:34:47.509094 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:34:47 crc kubenswrapper[4980]: I1206 03:34:47.509118 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:34:47 crc kubenswrapper[4980]: I1206 03:34:47.509131 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:34:47Z","lastTransitionTime":"2025-12-06T03:34:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:34:47 crc kubenswrapper[4980]: I1206 03:34:47.611796 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:34:47 crc kubenswrapper[4980]: I1206 03:34:47.611839 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:34:47 crc kubenswrapper[4980]: I1206 03:34:47.611851 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:34:47 crc kubenswrapper[4980]: I1206 03:34:47.611867 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:34:47 crc kubenswrapper[4980]: I1206 03:34:47.611876 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:34:47Z","lastTransitionTime":"2025-12-06T03:34:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:34:47 crc kubenswrapper[4980]: I1206 03:34:47.715447 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:34:47 crc kubenswrapper[4980]: I1206 03:34:47.715581 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:34:47 crc kubenswrapper[4980]: I1206 03:34:47.715614 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:34:47 crc kubenswrapper[4980]: I1206 03:34:47.715642 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:34:47 crc kubenswrapper[4980]: I1206 03:34:47.715663 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:34:47Z","lastTransitionTime":"2025-12-06T03:34:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:34:47 crc kubenswrapper[4980]: I1206 03:34:47.818541 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:34:47 crc kubenswrapper[4980]: I1206 03:34:47.818591 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:34:47 crc kubenswrapper[4980]: I1206 03:34:47.818603 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:34:47 crc kubenswrapper[4980]: I1206 03:34:47.818620 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:34:47 crc kubenswrapper[4980]: I1206 03:34:47.818633 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:34:47Z","lastTransitionTime":"2025-12-06T03:34:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:34:47 crc kubenswrapper[4980]: I1206 03:34:47.921493 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:34:47 crc kubenswrapper[4980]: I1206 03:34:47.921549 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:34:47 crc kubenswrapper[4980]: I1206 03:34:47.921559 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:34:47 crc kubenswrapper[4980]: I1206 03:34:47.921575 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:34:47 crc kubenswrapper[4980]: I1206 03:34:47.921586 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:34:47Z","lastTransitionTime":"2025-12-06T03:34:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:34:48 crc kubenswrapper[4980]: I1206 03:34:48.024239 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:34:48 crc kubenswrapper[4980]: I1206 03:34:48.024321 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:34:48 crc kubenswrapper[4980]: I1206 03:34:48.024332 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:34:48 crc kubenswrapper[4980]: I1206 03:34:48.024355 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:34:48 crc kubenswrapper[4980]: I1206 03:34:48.024368 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:34:48Z","lastTransitionTime":"2025-12-06T03:34:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:34:48 crc kubenswrapper[4980]: I1206 03:34:48.127054 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:34:48 crc kubenswrapper[4980]: I1206 03:34:48.127087 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:34:48 crc kubenswrapper[4980]: I1206 03:34:48.127098 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:34:48 crc kubenswrapper[4980]: I1206 03:34:48.127111 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:34:48 crc kubenswrapper[4980]: I1206 03:34:48.127120 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:34:48Z","lastTransitionTime":"2025-12-06T03:34:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:34:48 crc kubenswrapper[4980]: I1206 03:34:48.184257 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 03:34:48 crc kubenswrapper[4980]: I1206 03:34:48.184342 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 03:34:48 crc kubenswrapper[4980]: E1206 03:34:48.184450 4980 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 03:34:48 crc kubenswrapper[4980]: E1206 03:34:48.184615 4980 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 03:34:48 crc kubenswrapper[4980]: I1206 03:34:48.184784 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-mcprh" Dec 06 03:34:48 crc kubenswrapper[4980]: E1206 03:34:48.185007 4980 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-mcprh" podUID="e6e88c5e-a255-449a-ae38-ac4d73a8e984" Dec 06 03:34:48 crc kubenswrapper[4980]: I1206 03:34:48.229995 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:34:48 crc kubenswrapper[4980]: I1206 03:34:48.230024 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:34:48 crc kubenswrapper[4980]: I1206 03:34:48.230032 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:34:48 crc kubenswrapper[4980]: I1206 03:34:48.230043 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:34:48 crc kubenswrapper[4980]: I1206 03:34:48.230052 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:34:48Z","lastTransitionTime":"2025-12-06T03:34:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:34:48 crc kubenswrapper[4980]: I1206 03:34:48.333190 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:34:48 crc kubenswrapper[4980]: I1206 03:34:48.333263 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:34:48 crc kubenswrapper[4980]: I1206 03:34:48.333282 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:34:48 crc kubenswrapper[4980]: I1206 03:34:48.333308 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:34:48 crc kubenswrapper[4980]: I1206 03:34:48.333324 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:34:48Z","lastTransitionTime":"2025-12-06T03:34:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:34:48 crc kubenswrapper[4980]: I1206 03:34:48.441838 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:34:48 crc kubenswrapper[4980]: I1206 03:34:48.441911 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:34:48 crc kubenswrapper[4980]: I1206 03:34:48.441929 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:34:48 crc kubenswrapper[4980]: I1206 03:34:48.441956 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:34:48 crc kubenswrapper[4980]: I1206 03:34:48.441976 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:34:48Z","lastTransitionTime":"2025-12-06T03:34:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:34:48 crc kubenswrapper[4980]: I1206 03:34:48.545055 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:34:48 crc kubenswrapper[4980]: I1206 03:34:48.545128 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:34:48 crc kubenswrapper[4980]: I1206 03:34:48.545146 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:34:48 crc kubenswrapper[4980]: I1206 03:34:48.545170 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:34:48 crc kubenswrapper[4980]: I1206 03:34:48.545191 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:34:48Z","lastTransitionTime":"2025-12-06T03:34:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:34:48 crc kubenswrapper[4980]: I1206 03:34:48.648946 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:34:48 crc kubenswrapper[4980]: I1206 03:34:48.649028 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:34:48 crc kubenswrapper[4980]: I1206 03:34:48.649047 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:34:48 crc kubenswrapper[4980]: I1206 03:34:48.649072 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:34:48 crc kubenswrapper[4980]: I1206 03:34:48.649090 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:34:48Z","lastTransitionTime":"2025-12-06T03:34:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:34:48 crc kubenswrapper[4980]: I1206 03:34:48.712889 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:34:48 crc kubenswrapper[4980]: I1206 03:34:48.713001 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:34:48 crc kubenswrapper[4980]: I1206 03:34:48.713021 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:34:48 crc kubenswrapper[4980]: I1206 03:34:48.713107 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:34:48 crc kubenswrapper[4980]: I1206 03:34:48.713246 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:34:48Z","lastTransitionTime":"2025-12-06T03:34:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:34:48 crc kubenswrapper[4980]: E1206 03:34:48.734554 4980 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T03:34:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:48Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T03:34:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:48Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T03:34:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:48Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T03:34:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:48Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"9b4ed4f8-ed8d-447b-9e1f-13defbdcf0f7\\\",\\\"systemUUID\\\":\\\"a5b493aa-69de-4688-8ff6-1e5f49b7f014\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:48Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:48 crc kubenswrapper[4980]: I1206 03:34:48.739622 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:34:48 crc kubenswrapper[4980]: I1206 03:34:48.739690 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:34:48 crc kubenswrapper[4980]: I1206 03:34:48.739754 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:34:48 crc kubenswrapper[4980]: I1206 03:34:48.739803 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:34:48 crc kubenswrapper[4980]: I1206 03:34:48.739827 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:34:48Z","lastTransitionTime":"2025-12-06T03:34:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:34:48 crc kubenswrapper[4980]: E1206 03:34:48.758026 4980 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T03:34:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:48Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T03:34:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:48Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T03:34:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:48Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T03:34:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:48Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"9b4ed4f8-ed8d-447b-9e1f-13defbdcf0f7\\\",\\\"systemUUID\\\":\\\"a5b493aa-69de-4688-8ff6-1e5f49b7f014\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:48Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:48 crc kubenswrapper[4980]: I1206 03:34:48.763198 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:34:48 crc kubenswrapper[4980]: I1206 03:34:48.763254 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:34:48 crc kubenswrapper[4980]: I1206 03:34:48.763269 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:34:48 crc kubenswrapper[4980]: I1206 03:34:48.763284 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:34:48 crc kubenswrapper[4980]: I1206 03:34:48.763296 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:34:48Z","lastTransitionTime":"2025-12-06T03:34:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:34:48 crc kubenswrapper[4980]: E1206 03:34:48.780268 4980 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T03:34:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:48Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T03:34:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:48Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T03:34:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:48Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T03:34:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:48Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"9b4ed4f8-ed8d-447b-9e1f-13defbdcf0f7\\\",\\\"systemUUID\\\":\\\"a5b493aa-69de-4688-8ff6-1e5f49b7f014\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:48Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:48 crc kubenswrapper[4980]: I1206 03:34:48.785973 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:34:48 crc kubenswrapper[4980]: I1206 03:34:48.786047 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:34:48 crc kubenswrapper[4980]: I1206 03:34:48.786070 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:34:48 crc kubenswrapper[4980]: I1206 03:34:48.786101 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:34:48 crc kubenswrapper[4980]: I1206 03:34:48.786122 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:34:48Z","lastTransitionTime":"2025-12-06T03:34:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:34:48 crc kubenswrapper[4980]: E1206 03:34:48.806505 4980 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T03:34:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:48Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T03:34:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:48Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T03:34:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:48Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T03:34:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:48Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"9b4ed4f8-ed8d-447b-9e1f-13defbdcf0f7\\\",\\\"systemUUID\\\":\\\"a5b493aa-69de-4688-8ff6-1e5f49b7f014\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:48Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:48 crc kubenswrapper[4980]: I1206 03:34:48.812018 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:34:48 crc kubenswrapper[4980]: I1206 03:34:48.812095 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:34:48 crc kubenswrapper[4980]: I1206 03:34:48.812119 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:34:48 crc kubenswrapper[4980]: I1206 03:34:48.812139 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:34:48 crc kubenswrapper[4980]: I1206 03:34:48.812156 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:34:48Z","lastTransitionTime":"2025-12-06T03:34:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:34:48 crc kubenswrapper[4980]: E1206 03:34:48.832363 4980 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T03:34:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:48Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T03:34:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:48Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T03:34:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:48Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T03:34:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:48Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"9b4ed4f8-ed8d-447b-9e1f-13defbdcf0f7\\\",\\\"systemUUID\\\":\\\"a5b493aa-69de-4688-8ff6-1e5f49b7f014\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:48Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:48 crc kubenswrapper[4980]: E1206 03:34:48.832644 4980 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 06 03:34:48 crc kubenswrapper[4980]: I1206 03:34:48.834402 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:34:48 crc kubenswrapper[4980]: I1206 03:34:48.834477 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:34:48 crc kubenswrapper[4980]: I1206 03:34:48.834509 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:34:48 crc kubenswrapper[4980]: I1206 03:34:48.834577 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:34:48 crc kubenswrapper[4980]: I1206 03:34:48.834613 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:34:48Z","lastTransitionTime":"2025-12-06T03:34:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:34:48 crc kubenswrapper[4980]: I1206 03:34:48.937611 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:34:48 crc kubenswrapper[4980]: I1206 03:34:48.937681 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:34:48 crc kubenswrapper[4980]: I1206 03:34:48.937700 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:34:48 crc kubenswrapper[4980]: I1206 03:34:48.937723 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:34:48 crc kubenswrapper[4980]: I1206 03:34:48.937736 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:34:48Z","lastTransitionTime":"2025-12-06T03:34:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:34:49 crc kubenswrapper[4980]: I1206 03:34:49.041698 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:34:49 crc kubenswrapper[4980]: I1206 03:34:49.041760 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:34:49 crc kubenswrapper[4980]: I1206 03:34:49.041779 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:34:49 crc kubenswrapper[4980]: I1206 03:34:49.041803 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:34:49 crc kubenswrapper[4980]: I1206 03:34:49.041822 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:34:49Z","lastTransitionTime":"2025-12-06T03:34:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:34:49 crc kubenswrapper[4980]: I1206 03:34:49.144318 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:34:49 crc kubenswrapper[4980]: I1206 03:34:49.144357 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:34:49 crc kubenswrapper[4980]: I1206 03:34:49.144366 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:34:49 crc kubenswrapper[4980]: I1206 03:34:49.144379 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:34:49 crc kubenswrapper[4980]: I1206 03:34:49.144391 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:34:49Z","lastTransitionTime":"2025-12-06T03:34:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:34:49 crc kubenswrapper[4980]: I1206 03:34:49.183934 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 03:34:49 crc kubenswrapper[4980]: E1206 03:34:49.184075 4980 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 03:34:49 crc kubenswrapper[4980]: I1206 03:34:49.246876 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:34:49 crc kubenswrapper[4980]: I1206 03:34:49.246943 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:34:49 crc kubenswrapper[4980]: I1206 03:34:49.246960 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:34:49 crc kubenswrapper[4980]: I1206 03:34:49.246986 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:34:49 crc kubenswrapper[4980]: I1206 03:34:49.247007 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:34:49Z","lastTransitionTime":"2025-12-06T03:34:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:34:49 crc kubenswrapper[4980]: I1206 03:34:49.349502 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:34:49 crc kubenswrapper[4980]: I1206 03:34:49.349578 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:34:49 crc kubenswrapper[4980]: I1206 03:34:49.349591 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:34:49 crc kubenswrapper[4980]: I1206 03:34:49.349608 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:34:49 crc kubenswrapper[4980]: I1206 03:34:49.349621 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:34:49Z","lastTransitionTime":"2025-12-06T03:34:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:34:49 crc kubenswrapper[4980]: I1206 03:34:49.453299 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:34:49 crc kubenswrapper[4980]: I1206 03:34:49.453388 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:34:49 crc kubenswrapper[4980]: I1206 03:34:49.453410 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:34:49 crc kubenswrapper[4980]: I1206 03:34:49.453442 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:34:49 crc kubenswrapper[4980]: I1206 03:34:49.453467 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:34:49Z","lastTransitionTime":"2025-12-06T03:34:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:34:49 crc kubenswrapper[4980]: I1206 03:34:49.556922 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:34:49 crc kubenswrapper[4980]: I1206 03:34:49.556961 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:34:49 crc kubenswrapper[4980]: I1206 03:34:49.556971 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:34:49 crc kubenswrapper[4980]: I1206 03:34:49.556985 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:34:49 crc kubenswrapper[4980]: I1206 03:34:49.556994 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:34:49Z","lastTransitionTime":"2025-12-06T03:34:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:34:49 crc kubenswrapper[4980]: I1206 03:34:49.661093 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:34:49 crc kubenswrapper[4980]: I1206 03:34:49.661142 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:34:49 crc kubenswrapper[4980]: I1206 03:34:49.661154 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:34:49 crc kubenswrapper[4980]: I1206 03:34:49.661175 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:34:49 crc kubenswrapper[4980]: I1206 03:34:49.661188 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:34:49Z","lastTransitionTime":"2025-12-06T03:34:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:34:49 crc kubenswrapper[4980]: I1206 03:34:49.764083 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:34:49 crc kubenswrapper[4980]: I1206 03:34:49.764151 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:34:49 crc kubenswrapper[4980]: I1206 03:34:49.764175 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:34:49 crc kubenswrapper[4980]: I1206 03:34:49.764209 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:34:49 crc kubenswrapper[4980]: I1206 03:34:49.764234 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:34:49Z","lastTransitionTime":"2025-12-06T03:34:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:34:49 crc kubenswrapper[4980]: I1206 03:34:49.867913 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:34:49 crc kubenswrapper[4980]: I1206 03:34:49.868021 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:34:49 crc kubenswrapper[4980]: I1206 03:34:49.868095 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:34:49 crc kubenswrapper[4980]: I1206 03:34:49.868129 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:34:49 crc kubenswrapper[4980]: I1206 03:34:49.868220 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:34:49Z","lastTransitionTime":"2025-12-06T03:34:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:34:49 crc kubenswrapper[4980]: I1206 03:34:49.971970 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:34:49 crc kubenswrapper[4980]: I1206 03:34:49.972021 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:34:49 crc kubenswrapper[4980]: I1206 03:34:49.972033 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:34:49 crc kubenswrapper[4980]: I1206 03:34:49.972049 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:34:49 crc kubenswrapper[4980]: I1206 03:34:49.972058 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:34:49Z","lastTransitionTime":"2025-12-06T03:34:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:34:50 crc kubenswrapper[4980]: I1206 03:34:50.075370 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:34:50 crc kubenswrapper[4980]: I1206 03:34:50.075429 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:34:50 crc kubenswrapper[4980]: I1206 03:34:50.075445 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:34:50 crc kubenswrapper[4980]: I1206 03:34:50.075468 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:34:50 crc kubenswrapper[4980]: I1206 03:34:50.075484 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:34:50Z","lastTransitionTime":"2025-12-06T03:34:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:34:50 crc kubenswrapper[4980]: I1206 03:34:50.178968 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:34:50 crc kubenswrapper[4980]: I1206 03:34:50.179007 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:34:50 crc kubenswrapper[4980]: I1206 03:34:50.179017 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:34:50 crc kubenswrapper[4980]: I1206 03:34:50.179032 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:34:50 crc kubenswrapper[4980]: I1206 03:34:50.179043 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:34:50Z","lastTransitionTime":"2025-12-06T03:34:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:34:50 crc kubenswrapper[4980]: I1206 03:34:50.183930 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 03:34:50 crc kubenswrapper[4980]: E1206 03:34:50.184036 4980 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 03:34:50 crc kubenswrapper[4980]: I1206 03:34:50.184191 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-mcprh" Dec 06 03:34:50 crc kubenswrapper[4980]: E1206 03:34:50.184252 4980 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-mcprh" podUID="e6e88c5e-a255-449a-ae38-ac4d73a8e984" Dec 06 03:34:50 crc kubenswrapper[4980]: I1206 03:34:50.184508 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 03:34:50 crc kubenswrapper[4980]: E1206 03:34:50.184568 4980 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 03:34:50 crc kubenswrapper[4980]: I1206 03:34:50.281983 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:34:50 crc kubenswrapper[4980]: I1206 03:34:50.282041 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:34:50 crc kubenswrapper[4980]: I1206 03:34:50.282056 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:34:50 crc kubenswrapper[4980]: I1206 03:34:50.282076 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:34:50 crc kubenswrapper[4980]: I1206 03:34:50.282088 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:34:50Z","lastTransitionTime":"2025-12-06T03:34:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:34:50 crc kubenswrapper[4980]: I1206 03:34:50.384821 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:34:50 crc kubenswrapper[4980]: I1206 03:34:50.384862 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:34:50 crc kubenswrapper[4980]: I1206 03:34:50.384874 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:34:50 crc kubenswrapper[4980]: I1206 03:34:50.384895 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:34:50 crc kubenswrapper[4980]: I1206 03:34:50.384906 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:34:50Z","lastTransitionTime":"2025-12-06T03:34:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:34:50 crc kubenswrapper[4980]: I1206 03:34:50.487684 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:34:50 crc kubenswrapper[4980]: I1206 03:34:50.487752 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:34:50 crc kubenswrapper[4980]: I1206 03:34:50.487765 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:34:50 crc kubenswrapper[4980]: I1206 03:34:50.487787 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:34:50 crc kubenswrapper[4980]: I1206 03:34:50.487803 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:34:50Z","lastTransitionTime":"2025-12-06T03:34:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:34:50 crc kubenswrapper[4980]: I1206 03:34:50.590924 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:34:50 crc kubenswrapper[4980]: I1206 03:34:50.590972 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:34:50 crc kubenswrapper[4980]: I1206 03:34:50.590988 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:34:50 crc kubenswrapper[4980]: I1206 03:34:50.591008 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:34:50 crc kubenswrapper[4980]: I1206 03:34:50.591023 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:34:50Z","lastTransitionTime":"2025-12-06T03:34:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:34:50 crc kubenswrapper[4980]: I1206 03:34:50.694601 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:34:50 crc kubenswrapper[4980]: I1206 03:34:50.694654 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:34:50 crc kubenswrapper[4980]: I1206 03:34:50.694667 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:34:50 crc kubenswrapper[4980]: I1206 03:34:50.694690 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:34:50 crc kubenswrapper[4980]: I1206 03:34:50.694705 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:34:50Z","lastTransitionTime":"2025-12-06T03:34:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:34:50 crc kubenswrapper[4980]: I1206 03:34:50.798430 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:34:50 crc kubenswrapper[4980]: I1206 03:34:50.798497 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:34:50 crc kubenswrapper[4980]: I1206 03:34:50.798545 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:34:50 crc kubenswrapper[4980]: I1206 03:34:50.798575 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:34:50 crc kubenswrapper[4980]: I1206 03:34:50.798594 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:34:50Z","lastTransitionTime":"2025-12-06T03:34:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:34:50 crc kubenswrapper[4980]: I1206 03:34:50.902564 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:34:50 crc kubenswrapper[4980]: I1206 03:34:50.902628 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:34:50 crc kubenswrapper[4980]: I1206 03:34:50.902646 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:34:50 crc kubenswrapper[4980]: I1206 03:34:50.902671 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:34:50 crc kubenswrapper[4980]: I1206 03:34:50.902688 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:34:50Z","lastTransitionTime":"2025-12-06T03:34:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:34:51 crc kubenswrapper[4980]: I1206 03:34:51.006169 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:34:51 crc kubenswrapper[4980]: I1206 03:34:51.006251 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:34:51 crc kubenswrapper[4980]: I1206 03:34:51.006269 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:34:51 crc kubenswrapper[4980]: I1206 03:34:51.006288 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:34:51 crc kubenswrapper[4980]: I1206 03:34:51.006300 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:34:51Z","lastTransitionTime":"2025-12-06T03:34:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:34:51 crc kubenswrapper[4980]: I1206 03:34:51.109809 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:34:51 crc kubenswrapper[4980]: I1206 03:34:51.109872 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:34:51 crc kubenswrapper[4980]: I1206 03:34:51.109894 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:34:51 crc kubenswrapper[4980]: I1206 03:34:51.109924 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:34:51 crc kubenswrapper[4980]: I1206 03:34:51.109945 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:34:51Z","lastTransitionTime":"2025-12-06T03:34:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:34:51 crc kubenswrapper[4980]: I1206 03:34:51.184595 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 03:34:51 crc kubenswrapper[4980]: E1206 03:34:51.184838 4980 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 03:34:51 crc kubenswrapper[4980]: I1206 03:34:51.207930 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7f958150-549d-4c80-8a7e-23c39e4f9500\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c28b6895b14e73f4f3eb7a8ce93076c05092ba9b58ea0e3e879c65bcfa227a21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://76cff8a2d652df38b04eefc053832baf4cbb5db778ff5e525b18c2825141b118\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://206f6f19fe3197f5d6798cc86c5473bf2decb02a2860ee7ffb5efb5f60952e8c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bafaa9670b05671fb57b658c2960bf3ee32c84e52f86a3a3bbab01d5a38ff0c4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://773a0bc2aab73ffc2717e987543defc9b786cc22300aa54b4ba4f3454504e331\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-06T03:34:22Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1206 03:34:15.305330 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1206 03:34:15.307249 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1592846546/tls.crt::/tmp/serving-cert-1592846546/tls.key\\\\\\\"\\\\nI1206 03:34:21.814037 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1206 03:34:21.816756 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1206 03:34:21.816780 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1206 03:34:21.816806 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1206 03:34:21.816813 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1206 03:34:21.831301 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1206 03:34:21.831311 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1206 03:34:21.831324 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 03:34:21.831330 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 03:34:21.831335 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1206 03:34:21.831339 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1206 03:34:21.831342 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1206 03:34:21.831346 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1206 03:34:21.911269 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:04Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d40e0e7446f8416bf9da94fc53c01be937f9a51a11eb101157d2ddd4597720d2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:04Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2cba89f2f1bed96d59f183628d900d9b72c91eb472e273c8bed0babcab485910\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2cba89f2f1bed96d59f183628d900d9b72c91eb472e273c8bed0babcab485910\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:01Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:51Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:51 crc kubenswrapper[4980]: I1206 03:34:51.212681 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:34:51 crc kubenswrapper[4980]: I1206 03:34:51.212732 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:34:51 crc kubenswrapper[4980]: I1206 03:34:51.212744 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:34:51 crc kubenswrapper[4980]: I1206 03:34:51.212762 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:34:51 crc kubenswrapper[4980]: I1206 03:34:51.212774 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:34:51Z","lastTransitionTime":"2025-12-06T03:34:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:34:51 crc kubenswrapper[4980]: I1206 03:34:51.226209 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6ed68f66-1cf6-465a-8f66-76d5af6a44d8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://725e84f6bdfb722a8ce69158d03ea4c66a9f087642d55cb9651e003c81be0c8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://71f7b492ea07cdf7dc65fc48184ec1e64914eb1811bbfdddc0fb14c8ad6cb296\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b543aa4ee03d5a14909b67d876c5675c02e1df2a89a3c9ac8d3fb587fea53d8f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://32af3427e872a70cb3c2357116f04cabc8d3d9b69e7306b2e0de2ae22f9469d4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:01Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:51Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:51 crc kubenswrapper[4980]: I1206 03:34:51.242880 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:22Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:51Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:51 crc kubenswrapper[4980]: I1206 03:34:51.259605 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-sv5sz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"873977aa-4eb0-4526-bc0b-757ad2dea2a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01f0cc855c822a71ea8ef6a7ce6e25747ff9956691735973f935c4a28672a7a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lg7tx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:23Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-sv5sz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:51Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:51 crc kubenswrapper[4980]: I1206 03:34:51.282482 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5w4fr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5af0e768-3c9c-4401-ab56-57516cd3170e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:23Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:23Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://768b8e1fe1a00a8467d744da5ad3f61006b0a3612c96f8e4d0ee6fcb93195f9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfzrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6342903b318317ee8d84c2f5d467cb25c811a789be0d6ab0e0cba37b5611e16d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfzrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d620685574db7594132ae24a91ec0a93dce20a9a289540deebb39650b8a1816\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfzrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c67ec8f9c1764fcdf215d52822c2d4f0437dc7f056f05d708767121f6c1f3515\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfzrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f35beae5339eeaf26494ed678e3c1bd02349eae2a4b4615325b17923f3b653b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfzrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a940ec6ee523313728f7e8ae934b7e0b0cff82771e3e64723309db1fae013c3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfzrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6cb5cc9f6f22133cebf095db595e5062d799d7e3c8ca0bf31d3809a296e507cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6cb5cc9f6f22133cebf095db595e5062d799d7e3c8ca0bf31d3809a296e507cc\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-06T03:34:39Z\\\",\\\"message\\\":\\\"lector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1206 03:34:39.306579 6370 reflector.go:311] Stopping reflector *v1.AdminPolicyBasedExternalRoute (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/adminpolicybasedroute/v1/apis/informers/externalversions/factory.go:140\\\\nI1206 03:34:39.306665 6370 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1206 03:34:39.306723 6370 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI1206 03:34:39.306893 6370 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1206 03:34:39.307227 6370 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1206 03:34:39.307306 6370 reflector.go:311] Stopping reflector *v1.NetworkAttachmentDefinition (0s) from github.com/k8snetworkplumbingwg/network-attachment-definition-client/pkg/client/informers/externalversions/factory.go:117\\\\nI1206 03:34:39.307466 6370 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:37Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-5w4fr_openshift-ovn-kubernetes(5af0e768-3c9c-4401-ab56-57516cd3170e)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfzrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe261f4322b85702181d7473ab2a9a4d2d4b869d586c3091f811b12a81a63aae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfzrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d876df04b5ab0ba07a5cbcafbe4648403c5014c59eb7d40adee2657c3a255473\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d876df04b5ab0ba07a5cbcafbe4648403c5014c59eb7d40adee2657c3a255473\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfzrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:23Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-5w4fr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:51Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:51 crc kubenswrapper[4980]: I1206 03:34:51.300478 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:22Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:51Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:51 crc kubenswrapper[4980]: I1206 03:34:51.316384 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:34:51 crc kubenswrapper[4980]: I1206 03:34:51.316462 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:34:51 crc kubenswrapper[4980]: I1206 03:34:51.316475 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:34:51 crc kubenswrapper[4980]: I1206 03:34:51.316546 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:34:51 crc kubenswrapper[4980]: I1206 03:34:51.316563 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:34:51Z","lastTransitionTime":"2025-12-06T03:34:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:34:51 crc kubenswrapper[4980]: I1206 03:34:51.317622 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://334ea6a273fd7d268906c9f4a6dbf6f4df4fcdbf38f3d7d248cb7d6cf8537794\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ec44dcd5e4bf0a0466bb586ebd61259364dbbfe22499f4cbff854e3bd153082\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:51Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:51 crc kubenswrapper[4980]: I1206 03:34:51.335483 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-r5zfb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"320f44d1-a671-4a91-b328-a8b0fdd8f23a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://92393ecaaef8cfcfd01837671f8cc89803dd570b637b4a373b40b7f487788826\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zhm5t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52e8db644463794041978ced2c0591c13e0ada8e5f1a78a94a0a665218a37b5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zhm5t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:23Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-r5zfb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:51Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:51 crc kubenswrapper[4980]: I1206 03:34:51.351318 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-cvzg8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f1cde706-268d-4d26-818f-2f42bfb37b5f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://86f9a77acae360740be827eb0d1e1516e8a49cb9b197b6f5f9468ac70a101158\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rk86l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:29Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-cvzg8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:51Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:51 crc kubenswrapper[4980]: I1206 03:34:51.379335 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cab608b3-a02f-475b-9aae-9b8f5d878de9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f7fcc44b42bcbb55509048e3a566bb557333c4a83e33d7b0fba8eb514317909\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c22e7dd8886c2dfbb742900ed86ec7b0c2edf9a1363e74f6666961d928f06c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://473e792b2d0a9644854ee0581c6120fc9c989bc342899ef72b1d056080fd4694\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ee8e936dfc39361d6f99460910591ca8c901a20f0d274bbb6aaebfa8a00ee06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://77bc12768b018625ba11a6bfd66688f926c4b7f3bc1d0e41c6551a18b485ecf3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://27cf07cd96893900da4afecd0d56fd9269bc60c1d23246b4685fcd8756357c19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://27cf07cd96893900da4afecd0d56fd9269bc60c1d23246b4685fcd8756357c19\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c49feb24566a8284378670fa0390710d38c237acc94ffb654f90f3fa4626e4c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c49feb24566a8284378670fa0390710d38c237acc94ffb654f90f3fa4626e4c8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:03Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://bb74d1c3d752a379fa394c450217badb45511fbcafad62f7b618131cab94b29e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bb74d1c3d752a379fa394c450217badb45511fbcafad62f7b618131cab94b29e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:01Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:51Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:51 crc kubenswrapper[4980]: I1206 03:34:51.390130 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://502c727d6baa31f400bee14c34b19abbe38af873f28f6a82a43547431755d981\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:51Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:51 crc kubenswrapper[4980]: I1206 03:34:51.403072 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-dx44q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5bac22a-512c-49d0-90db-47de6aaaefd0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://97469dea2cd5393da0acf784ccc1969c8a4d07da7d51b14cdc955013ee9710ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f7mdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3b5dab476c7f227a6e4a2eae8469432667fb9fad345d9419ba59feab9fd385e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c3b5dab476c7f227a6e4a2eae8469432667fb9fad345d9419ba59feab9fd385e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f7mdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e921499231cbdf26b8918c98f9ad8c322fd7848865caeef048470de3e2c0efba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e921499231cbdf26b8918c98f9ad8c322fd7848865caeef048470de3e2c0efba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f7mdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7cbfe7d8cdb89916c238296592fab45b9e92e2c585baa6d67230decd1c84c5bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7cbfe7d8cdb89916c238296592fab45b9e92e2c585baa6d67230decd1c84c5bb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f7mdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f45cbf23b6b0d208cefc4b493f901bab9d71ccf2f1456507d54fb1f30c4fd48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7f45cbf23b6b0d208cefc4b493f901bab9d71ccf2f1456507d54fb1f30c4fd48\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f7mdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://344ce4b145ff021fcaf163803d8ad8cc6e9565bf7f0992e2653c43db663d8411\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://344ce4b145ff021fcaf163803d8ad8cc6e9565bf7f0992e2653c43db663d8411\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f7mdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://619e56a9937dd6376ab471ba16da30e5d77f443f5bd33610bea5ec85cd0446ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://619e56a9937dd6376ab471ba16da30e5d77f443f5bd33610bea5ec85cd0446ef\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f7mdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:23Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-dx44q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:51Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:51 crc kubenswrapper[4980]: I1206 03:34:51.414911 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-6kxfq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"951b6deb-bcea-4332-8ae8-c23ac7aaef36\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://423d30c3e65b04bdf6ff5ef53d8a7575a434c05e28489370a7370381ebc29de9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b8htf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://282df50dd9653c1cbe97fc2041d9e9ea91a992661682652aafc12e566d8f1b81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b8htf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:37Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-6kxfq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:51Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:51 crc kubenswrapper[4980]: I1206 03:34:51.419893 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:34:51 crc kubenswrapper[4980]: I1206 03:34:51.419925 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:34:51 crc kubenswrapper[4980]: I1206 03:34:51.419938 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:34:51 crc kubenswrapper[4980]: I1206 03:34:51.419958 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:34:51 crc kubenswrapper[4980]: I1206 03:34:51.419972 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:34:51Z","lastTransitionTime":"2025-12-06T03:34:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:34:51 crc kubenswrapper[4980]: I1206 03:34:51.430119 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-mcprh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e6e88c5e-a255-449a-ae38-ac4d73a8e984\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g7xz9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g7xz9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:38Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-mcprh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:51Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:51 crc kubenswrapper[4980]: I1206 03:34:51.444546 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:51Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:51 crc kubenswrapper[4980]: I1206 03:34:51.459309 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4ef01e7d9809fd9cfc46e3a58b07159cdc6f5a3c210d1c9efc49bca27b30858\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:51Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:51 crc kubenswrapper[4980]: I1206 03:34:51.473492 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-kq4z8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f96c1f6c-55e8-4d86-bff3-5381581b7022\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://799498681afa7653408bc23b1e82d75c16e3c809d2ee7f322780e20c537eb710\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bdx7g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:23Z\\\"}}\" for pod \"openshift-multus\"/\"multus-kq4z8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:51Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:51 crc kubenswrapper[4980]: I1206 03:34:51.522545 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:34:51 crc kubenswrapper[4980]: I1206 03:34:51.522593 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:34:51 crc kubenswrapper[4980]: I1206 03:34:51.522602 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:34:51 crc kubenswrapper[4980]: I1206 03:34:51.522615 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:34:51 crc kubenswrapper[4980]: I1206 03:34:51.522624 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:34:51Z","lastTransitionTime":"2025-12-06T03:34:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:34:51 crc kubenswrapper[4980]: I1206 03:34:51.625491 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:34:51 crc kubenswrapper[4980]: I1206 03:34:51.625560 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:34:51 crc kubenswrapper[4980]: I1206 03:34:51.625576 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:34:51 crc kubenswrapper[4980]: I1206 03:34:51.625597 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:34:51 crc kubenswrapper[4980]: I1206 03:34:51.625615 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:34:51Z","lastTransitionTime":"2025-12-06T03:34:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:34:51 crc kubenswrapper[4980]: I1206 03:34:51.728964 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:34:51 crc kubenswrapper[4980]: I1206 03:34:51.729024 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:34:51 crc kubenswrapper[4980]: I1206 03:34:51.729040 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:34:51 crc kubenswrapper[4980]: I1206 03:34:51.729076 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:34:51 crc kubenswrapper[4980]: I1206 03:34:51.729111 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:34:51Z","lastTransitionTime":"2025-12-06T03:34:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:34:51 crc kubenswrapper[4980]: I1206 03:34:51.832689 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:34:51 crc kubenswrapper[4980]: I1206 03:34:51.832739 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:34:51 crc kubenswrapper[4980]: I1206 03:34:51.832755 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:34:51 crc kubenswrapper[4980]: I1206 03:34:51.832776 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:34:51 crc kubenswrapper[4980]: I1206 03:34:51.832790 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:34:51Z","lastTransitionTime":"2025-12-06T03:34:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:34:51 crc kubenswrapper[4980]: I1206 03:34:51.936124 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:34:51 crc kubenswrapper[4980]: I1206 03:34:51.936176 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:34:51 crc kubenswrapper[4980]: I1206 03:34:51.936188 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:34:51 crc kubenswrapper[4980]: I1206 03:34:51.936208 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:34:51 crc kubenswrapper[4980]: I1206 03:34:51.936222 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:34:51Z","lastTransitionTime":"2025-12-06T03:34:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:34:52 crc kubenswrapper[4980]: I1206 03:34:52.039016 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:34:52 crc kubenswrapper[4980]: I1206 03:34:52.040073 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:34:52 crc kubenswrapper[4980]: I1206 03:34:52.040310 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:34:52 crc kubenswrapper[4980]: I1206 03:34:52.040567 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:34:52 crc kubenswrapper[4980]: I1206 03:34:52.040824 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:34:52Z","lastTransitionTime":"2025-12-06T03:34:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:34:52 crc kubenswrapper[4980]: I1206 03:34:52.145060 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:34:52 crc kubenswrapper[4980]: I1206 03:34:52.145141 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:34:52 crc kubenswrapper[4980]: I1206 03:34:52.145167 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:34:52 crc kubenswrapper[4980]: I1206 03:34:52.145203 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:34:52 crc kubenswrapper[4980]: I1206 03:34:52.145227 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:34:52Z","lastTransitionTime":"2025-12-06T03:34:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:34:52 crc kubenswrapper[4980]: I1206 03:34:52.184434 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 03:34:52 crc kubenswrapper[4980]: I1206 03:34:52.184636 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-mcprh" Dec 06 03:34:52 crc kubenswrapper[4980]: E1206 03:34:52.184935 4980 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 03:34:52 crc kubenswrapper[4980]: I1206 03:34:52.185210 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 03:34:52 crc kubenswrapper[4980]: E1206 03:34:52.185316 4980 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 03:34:52 crc kubenswrapper[4980]: E1206 03:34:52.185671 4980 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-mcprh" podUID="e6e88c5e-a255-449a-ae38-ac4d73a8e984" Dec 06 03:34:52 crc kubenswrapper[4980]: I1206 03:34:52.187059 4980 scope.go:117] "RemoveContainer" containerID="6cb5cc9f6f22133cebf095db595e5062d799d7e3c8ca0bf31d3809a296e507cc" Dec 06 03:34:52 crc kubenswrapper[4980]: I1206 03:34:52.247929 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:34:52 crc kubenswrapper[4980]: I1206 03:34:52.247985 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:34:52 crc kubenswrapper[4980]: I1206 03:34:52.248002 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:34:52 crc kubenswrapper[4980]: I1206 03:34:52.248021 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:34:52 crc kubenswrapper[4980]: I1206 03:34:52.248034 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:34:52Z","lastTransitionTime":"2025-12-06T03:34:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:34:52 crc kubenswrapper[4980]: I1206 03:34:52.351485 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:34:52 crc kubenswrapper[4980]: I1206 03:34:52.351575 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:34:52 crc kubenswrapper[4980]: I1206 03:34:52.351592 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:34:52 crc kubenswrapper[4980]: I1206 03:34:52.351625 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:34:52 crc kubenswrapper[4980]: I1206 03:34:52.351641 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:34:52Z","lastTransitionTime":"2025-12-06T03:34:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:34:52 crc kubenswrapper[4980]: I1206 03:34:52.457852 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:34:52 crc kubenswrapper[4980]: I1206 03:34:52.457991 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:34:52 crc kubenswrapper[4980]: I1206 03:34:52.458014 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:34:52 crc kubenswrapper[4980]: I1206 03:34:52.458115 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:34:52 crc kubenswrapper[4980]: I1206 03:34:52.458160 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:34:52Z","lastTransitionTime":"2025-12-06T03:34:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:34:52 crc kubenswrapper[4980]: I1206 03:34:52.586788 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:34:52 crc kubenswrapper[4980]: I1206 03:34:52.586843 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:34:52 crc kubenswrapper[4980]: I1206 03:34:52.586856 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:34:52 crc kubenswrapper[4980]: I1206 03:34:52.586916 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:34:52 crc kubenswrapper[4980]: I1206 03:34:52.586932 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:34:52Z","lastTransitionTime":"2025-12-06T03:34:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:34:52 crc kubenswrapper[4980]: I1206 03:34:52.690295 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:34:52 crc kubenswrapper[4980]: I1206 03:34:52.692987 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:34:52 crc kubenswrapper[4980]: I1206 03:34:52.693048 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:34:52 crc kubenswrapper[4980]: I1206 03:34:52.693092 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:34:52 crc kubenswrapper[4980]: I1206 03:34:52.693109 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:34:52Z","lastTransitionTime":"2025-12-06T03:34:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:34:52 crc kubenswrapper[4980]: I1206 03:34:52.795717 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:34:52 crc kubenswrapper[4980]: I1206 03:34:52.795765 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:34:52 crc kubenswrapper[4980]: I1206 03:34:52.795776 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:34:52 crc kubenswrapper[4980]: I1206 03:34:52.795794 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:34:52 crc kubenswrapper[4980]: I1206 03:34:52.795808 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:34:52Z","lastTransitionTime":"2025-12-06T03:34:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:34:52 crc kubenswrapper[4980]: I1206 03:34:52.927204 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:34:52 crc kubenswrapper[4980]: I1206 03:34:52.927268 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:34:52 crc kubenswrapper[4980]: I1206 03:34:52.927281 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:34:52 crc kubenswrapper[4980]: I1206 03:34:52.927300 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:34:52 crc kubenswrapper[4980]: I1206 03:34:52.927313 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:34:52Z","lastTransitionTime":"2025-12-06T03:34:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:34:53 crc kubenswrapper[4980]: I1206 03:34:53.002713 4980 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-5w4fr_5af0e768-3c9c-4401-ab56-57516cd3170e/ovnkube-controller/1.log" Dec 06 03:34:53 crc kubenswrapper[4980]: I1206 03:34:53.007037 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5w4fr" event={"ID":"5af0e768-3c9c-4401-ab56-57516cd3170e","Type":"ContainerStarted","Data":"fae428d1a6eaeb2e6ddd4a5b1a1170e230635739ea0b39743e565e47fb7ee6e4"} Dec 06 03:34:53 crc kubenswrapper[4980]: I1206 03:34:53.008888 4980 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-5w4fr" Dec 06 03:34:53 crc kubenswrapper[4980]: I1206 03:34:53.027571 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7f958150-549d-4c80-8a7e-23c39e4f9500\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c28b6895b14e73f4f3eb7a8ce93076c05092ba9b58ea0e3e879c65bcfa227a21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://76cff8a2d652df38b04eefc053832baf4cbb5db778ff5e525b18c2825141b118\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://206f6f19fe3197f5d6798cc86c5473bf2decb02a2860ee7ffb5efb5f60952e8c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bafaa9670b05671fb57b658c2960bf3ee32c84e52f86a3a3bbab01d5a38ff0c4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://773a0bc2aab73ffc2717e987543defc9b786cc22300aa54b4ba4f3454504e331\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-06T03:34:22Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1206 03:34:15.305330 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1206 03:34:15.307249 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1592846546/tls.crt::/tmp/serving-cert-1592846546/tls.key\\\\\\\"\\\\nI1206 03:34:21.814037 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1206 03:34:21.816756 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1206 03:34:21.816780 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1206 03:34:21.816806 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1206 03:34:21.816813 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1206 03:34:21.831301 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1206 03:34:21.831311 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1206 03:34:21.831324 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 03:34:21.831330 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 03:34:21.831335 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1206 03:34:21.831339 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1206 03:34:21.831342 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1206 03:34:21.831346 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1206 03:34:21.911269 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:04Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d40e0e7446f8416bf9da94fc53c01be937f9a51a11eb101157d2ddd4597720d2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:04Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2cba89f2f1bed96d59f183628d900d9b72c91eb472e273c8bed0babcab485910\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2cba89f2f1bed96d59f183628d900d9b72c91eb472e273c8bed0babcab485910\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:01Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:53Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:53 crc kubenswrapper[4980]: I1206 03:34:53.030380 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:34:53 crc kubenswrapper[4980]: I1206 03:34:53.030454 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:34:53 crc kubenswrapper[4980]: I1206 03:34:53.030474 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:34:53 crc kubenswrapper[4980]: I1206 03:34:53.030501 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:34:53 crc kubenswrapper[4980]: I1206 03:34:53.030551 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:34:53Z","lastTransitionTime":"2025-12-06T03:34:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:34:53 crc kubenswrapper[4980]: I1206 03:34:53.046157 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6ed68f66-1cf6-465a-8f66-76d5af6a44d8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://725e84f6bdfb722a8ce69158d03ea4c66a9f087642d55cb9651e003c81be0c8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://71f7b492ea07cdf7dc65fc48184ec1e64914eb1811bbfdddc0fb14c8ad6cb296\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b543aa4ee03d5a14909b67d876c5675c02e1df2a89a3c9ac8d3fb587fea53d8f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://32af3427e872a70cb3c2357116f04cabc8d3d9b69e7306b2e0de2ae22f9469d4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:01Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:53Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:53 crc kubenswrapper[4980]: I1206 03:34:53.063431 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:22Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:53Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:53 crc kubenswrapper[4980]: I1206 03:34:53.082238 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://334ea6a273fd7d268906c9f4a6dbf6f4df4fcdbf38f3d7d248cb7d6cf8537794\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ec44dcd5e4bf0a0466bb586ebd61259364dbbfe22499f4cbff854e3bd153082\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:53Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:53 crc kubenswrapper[4980]: I1206 03:34:53.097109 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:22Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:53Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:53 crc kubenswrapper[4980]: I1206 03:34:53.112717 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-sv5sz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"873977aa-4eb0-4526-bc0b-757ad2dea2a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01f0cc855c822a71ea8ef6a7ce6e25747ff9956691735973f935c4a28672a7a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lg7tx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:23Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-sv5sz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:53Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:53 crc kubenswrapper[4980]: I1206 03:34:53.134647 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:34:53 crc kubenswrapper[4980]: I1206 03:34:53.134722 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:34:53 crc kubenswrapper[4980]: I1206 03:34:53.134740 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:34:53 crc kubenswrapper[4980]: I1206 03:34:53.134760 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:34:53 crc kubenswrapper[4980]: I1206 03:34:53.134776 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:34:53Z","lastTransitionTime":"2025-12-06T03:34:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:34:53 crc kubenswrapper[4980]: I1206 03:34:53.147070 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5w4fr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5af0e768-3c9c-4401-ab56-57516cd3170e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:23Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:23Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://768b8e1fe1a00a8467d744da5ad3f61006b0a3612c96f8e4d0ee6fcb93195f9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfzrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6342903b318317ee8d84c2f5d467cb25c811a789be0d6ab0e0cba37b5611e16d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfzrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d620685574db7594132ae24a91ec0a93dce20a9a289540deebb39650b8a1816\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfzrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c67ec8f9c1764fcdf215d52822c2d4f0437dc7f056f05d708767121f6c1f3515\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfzrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f35beae5339eeaf26494ed678e3c1bd02349eae2a4b4615325b17923f3b653b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfzrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a940ec6ee523313728f7e8ae934b7e0b0cff82771e3e64723309db1fae013c3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfzrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fae428d1a6eaeb2e6ddd4a5b1a1170e230635739ea0b39743e565e47fb7ee6e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6cb5cc9f6f22133cebf095db595e5062d799d7e3c8ca0bf31d3809a296e507cc\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-06T03:34:39Z\\\",\\\"message\\\":\\\"lector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1206 03:34:39.306579 6370 reflector.go:311] Stopping reflector *v1.AdminPolicyBasedExternalRoute (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/adminpolicybasedroute/v1/apis/informers/externalversions/factory.go:140\\\\nI1206 03:34:39.306665 6370 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1206 03:34:39.306723 6370 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI1206 03:34:39.306893 6370 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1206 03:34:39.307227 6370 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1206 03:34:39.307306 6370 reflector.go:311] Stopping reflector *v1.NetworkAttachmentDefinition (0s) from github.com/k8snetworkplumbingwg/network-attachment-definition-client/pkg/client/informers/externalversions/factory.go:117\\\\nI1206 03:34:39.307466 6370 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:37Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfzrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe261f4322b85702181d7473ab2a9a4d2d4b869d586c3091f811b12a81a63aae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfzrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d876df04b5ab0ba07a5cbcafbe4648403c5014c59eb7d40adee2657c3a255473\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d876df04b5ab0ba07a5cbcafbe4648403c5014c59eb7d40adee2657c3a255473\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfzrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:23Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-5w4fr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:53Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:53 crc kubenswrapper[4980]: I1206 03:34:53.176280 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cab608b3-a02f-475b-9aae-9b8f5d878de9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f7fcc44b42bcbb55509048e3a566bb557333c4a83e33d7b0fba8eb514317909\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c22e7dd8886c2dfbb742900ed86ec7b0c2edf9a1363e74f6666961d928f06c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://473e792b2d0a9644854ee0581c6120fc9c989bc342899ef72b1d056080fd4694\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ee8e936dfc39361d6f99460910591ca8c901a20f0d274bbb6aaebfa8a00ee06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://77bc12768b018625ba11a6bfd66688f926c4b7f3bc1d0e41c6551a18b485ecf3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://27cf07cd96893900da4afecd0d56fd9269bc60c1d23246b4685fcd8756357c19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://27cf07cd96893900da4afecd0d56fd9269bc60c1d23246b4685fcd8756357c19\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c49feb24566a8284378670fa0390710d38c237acc94ffb654f90f3fa4626e4c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c49feb24566a8284378670fa0390710d38c237acc94ffb654f90f3fa4626e4c8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:03Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://bb74d1c3d752a379fa394c450217badb45511fbcafad62f7b618131cab94b29e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bb74d1c3d752a379fa394c450217badb45511fbcafad62f7b618131cab94b29e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:01Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:53Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:53 crc kubenswrapper[4980]: I1206 03:34:53.183893 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 03:34:53 crc kubenswrapper[4980]: E1206 03:34:53.184056 4980 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 03:34:53 crc kubenswrapper[4980]: I1206 03:34:53.190856 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://502c727d6baa31f400bee14c34b19abbe38af873f28f6a82a43547431755d981\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:53Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:53 crc kubenswrapper[4980]: I1206 03:34:53.207129 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-dx44q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5bac22a-512c-49d0-90db-47de6aaaefd0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://97469dea2cd5393da0acf784ccc1969c8a4d07da7d51b14cdc955013ee9710ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f7mdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3b5dab476c7f227a6e4a2eae8469432667fb9fad345d9419ba59feab9fd385e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c3b5dab476c7f227a6e4a2eae8469432667fb9fad345d9419ba59feab9fd385e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f7mdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e921499231cbdf26b8918c98f9ad8c322fd7848865caeef048470de3e2c0efba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e921499231cbdf26b8918c98f9ad8c322fd7848865caeef048470de3e2c0efba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f7mdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7cbfe7d8cdb89916c238296592fab45b9e92e2c585baa6d67230decd1c84c5bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7cbfe7d8cdb89916c238296592fab45b9e92e2c585baa6d67230decd1c84c5bb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f7mdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f45cbf23b6b0d208cefc4b493f901bab9d71ccf2f1456507d54fb1f30c4fd48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7f45cbf23b6b0d208cefc4b493f901bab9d71ccf2f1456507d54fb1f30c4fd48\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f7mdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://344ce4b145ff021fcaf163803d8ad8cc6e9565bf7f0992e2653c43db663d8411\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://344ce4b145ff021fcaf163803d8ad8cc6e9565bf7f0992e2653c43db663d8411\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f7mdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://619e56a9937dd6376ab471ba16da30e5d77f443f5bd33610bea5ec85cd0446ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://619e56a9937dd6376ab471ba16da30e5d77f443f5bd33610bea5ec85cd0446ef\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f7mdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:23Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-dx44q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:53Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:53 crc kubenswrapper[4980]: I1206 03:34:53.221611 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-r5zfb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"320f44d1-a671-4a91-b328-a8b0fdd8f23a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://92393ecaaef8cfcfd01837671f8cc89803dd570b637b4a373b40b7f487788826\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zhm5t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52e8db644463794041978ced2c0591c13e0ada8e5f1a78a94a0a665218a37b5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zhm5t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:23Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-r5zfb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:53Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:53 crc kubenswrapper[4980]: I1206 03:34:53.239114 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-cvzg8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f1cde706-268d-4d26-818f-2f42bfb37b5f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://86f9a77acae360740be827eb0d1e1516e8a49cb9b197b6f5f9468ac70a101158\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rk86l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:29Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-cvzg8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:53Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:53 crc kubenswrapper[4980]: I1206 03:34:53.239170 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:34:53 crc kubenswrapper[4980]: I1206 03:34:53.239360 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:34:53 crc kubenswrapper[4980]: I1206 03:34:53.239371 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:34:53 crc kubenswrapper[4980]: I1206 03:34:53.239389 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:34:53 crc kubenswrapper[4980]: I1206 03:34:53.239400 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:34:53Z","lastTransitionTime":"2025-12-06T03:34:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:34:53 crc kubenswrapper[4980]: I1206 03:34:53.257695 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:53Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:53 crc kubenswrapper[4980]: I1206 03:34:53.276699 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4ef01e7d9809fd9cfc46e3a58b07159cdc6f5a3c210d1c9efc49bca27b30858\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:53Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:53 crc kubenswrapper[4980]: I1206 03:34:53.292703 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-kq4z8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f96c1f6c-55e8-4d86-bff3-5381581b7022\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://799498681afa7653408bc23b1e82d75c16e3c809d2ee7f322780e20c537eb710\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bdx7g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:23Z\\\"}}\" for pod \"openshift-multus\"/\"multus-kq4z8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:53Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:53 crc kubenswrapper[4980]: I1206 03:34:53.305717 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-6kxfq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"951b6deb-bcea-4332-8ae8-c23ac7aaef36\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://423d30c3e65b04bdf6ff5ef53d8a7575a434c05e28489370a7370381ebc29de9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b8htf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://282df50dd9653c1cbe97fc2041d9e9ea91a992661682652aafc12e566d8f1b81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b8htf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:37Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-6kxfq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:53Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:53 crc kubenswrapper[4980]: I1206 03:34:53.319554 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-mcprh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e6e88c5e-a255-449a-ae38-ac4d73a8e984\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g7xz9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g7xz9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:38Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-mcprh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:53Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:53 crc kubenswrapper[4980]: I1206 03:34:53.342486 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:34:53 crc kubenswrapper[4980]: I1206 03:34:53.343011 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:34:53 crc kubenswrapper[4980]: I1206 03:34:53.343031 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:34:53 crc kubenswrapper[4980]: I1206 03:34:53.343055 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:34:53 crc kubenswrapper[4980]: I1206 03:34:53.343069 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:34:53Z","lastTransitionTime":"2025-12-06T03:34:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:34:53 crc kubenswrapper[4980]: I1206 03:34:53.462102 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:34:53 crc kubenswrapper[4980]: I1206 03:34:53.462169 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:34:53 crc kubenswrapper[4980]: I1206 03:34:53.462192 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:34:53 crc kubenswrapper[4980]: I1206 03:34:53.462217 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:34:53 crc kubenswrapper[4980]: I1206 03:34:53.462230 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:34:53Z","lastTransitionTime":"2025-12-06T03:34:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:34:53 crc kubenswrapper[4980]: I1206 03:34:53.564233 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:34:53 crc kubenswrapper[4980]: I1206 03:34:53.564281 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:34:53 crc kubenswrapper[4980]: I1206 03:34:53.564290 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:34:53 crc kubenswrapper[4980]: I1206 03:34:53.564318 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:34:53 crc kubenswrapper[4980]: I1206 03:34:53.564329 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:34:53Z","lastTransitionTime":"2025-12-06T03:34:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:34:53 crc kubenswrapper[4980]: I1206 03:34:53.667936 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:34:53 crc kubenswrapper[4980]: I1206 03:34:53.668020 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:34:53 crc kubenswrapper[4980]: I1206 03:34:53.668040 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:34:53 crc kubenswrapper[4980]: I1206 03:34:53.668080 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:34:53 crc kubenswrapper[4980]: I1206 03:34:53.668107 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:34:53Z","lastTransitionTime":"2025-12-06T03:34:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:34:53 crc kubenswrapper[4980]: I1206 03:34:53.771417 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:34:53 crc kubenswrapper[4980]: I1206 03:34:53.771496 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:34:53 crc kubenswrapper[4980]: I1206 03:34:53.771523 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:34:53 crc kubenswrapper[4980]: I1206 03:34:53.771542 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:34:53 crc kubenswrapper[4980]: I1206 03:34:53.771554 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:34:53Z","lastTransitionTime":"2025-12-06T03:34:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:34:53 crc kubenswrapper[4980]: I1206 03:34:53.874601 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:34:53 crc kubenswrapper[4980]: I1206 03:34:53.874660 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:34:53 crc kubenswrapper[4980]: I1206 03:34:53.874672 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:34:53 crc kubenswrapper[4980]: I1206 03:34:53.874692 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:34:53 crc kubenswrapper[4980]: I1206 03:34:53.874704 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:34:53Z","lastTransitionTime":"2025-12-06T03:34:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:34:53 crc kubenswrapper[4980]: I1206 03:34:53.983262 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:34:53 crc kubenswrapper[4980]: I1206 03:34:53.983538 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:34:53 crc kubenswrapper[4980]: I1206 03:34:53.983586 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:34:53 crc kubenswrapper[4980]: I1206 03:34:53.983654 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:34:53 crc kubenswrapper[4980]: I1206 03:34:53.983704 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:34:53Z","lastTransitionTime":"2025-12-06T03:34:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:34:54 crc kubenswrapper[4980]: I1206 03:34:54.013385 4980 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-5w4fr_5af0e768-3c9c-4401-ab56-57516cd3170e/ovnkube-controller/2.log" Dec 06 03:34:54 crc kubenswrapper[4980]: I1206 03:34:54.014404 4980 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-5w4fr_5af0e768-3c9c-4401-ab56-57516cd3170e/ovnkube-controller/1.log" Dec 06 03:34:54 crc kubenswrapper[4980]: I1206 03:34:54.018839 4980 generic.go:334] "Generic (PLEG): container finished" podID="5af0e768-3c9c-4401-ab56-57516cd3170e" containerID="fae428d1a6eaeb2e6ddd4a5b1a1170e230635739ea0b39743e565e47fb7ee6e4" exitCode=1 Dec 06 03:34:54 crc kubenswrapper[4980]: I1206 03:34:54.018923 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5w4fr" event={"ID":"5af0e768-3c9c-4401-ab56-57516cd3170e","Type":"ContainerDied","Data":"fae428d1a6eaeb2e6ddd4a5b1a1170e230635739ea0b39743e565e47fb7ee6e4"} Dec 06 03:34:54 crc kubenswrapper[4980]: I1206 03:34:54.019020 4980 scope.go:117] "RemoveContainer" containerID="6cb5cc9f6f22133cebf095db595e5062d799d7e3c8ca0bf31d3809a296e507cc" Dec 06 03:34:54 crc kubenswrapper[4980]: I1206 03:34:54.020441 4980 scope.go:117] "RemoveContainer" containerID="fae428d1a6eaeb2e6ddd4a5b1a1170e230635739ea0b39743e565e47fb7ee6e4" Dec 06 03:34:54 crc kubenswrapper[4980]: E1206 03:34:54.020796 4980 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-5w4fr_openshift-ovn-kubernetes(5af0e768-3c9c-4401-ab56-57516cd3170e)\"" pod="openshift-ovn-kubernetes/ovnkube-node-5w4fr" podUID="5af0e768-3c9c-4401-ab56-57516cd3170e" Dec 06 03:34:54 crc kubenswrapper[4980]: I1206 03:34:54.042151 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6ed68f66-1cf6-465a-8f66-76d5af6a44d8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://725e84f6bdfb722a8ce69158d03ea4c66a9f087642d55cb9651e003c81be0c8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://71f7b492ea07cdf7dc65fc48184ec1e64914eb1811bbfdddc0fb14c8ad6cb296\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b543aa4ee03d5a14909b67d876c5675c02e1df2a89a3c9ac8d3fb587fea53d8f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://32af3427e872a70cb3c2357116f04cabc8d3d9b69e7306b2e0de2ae22f9469d4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:01Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:54Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:54 crc kubenswrapper[4980]: I1206 03:34:54.058536 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7f958150-549d-4c80-8a7e-23c39e4f9500\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c28b6895b14e73f4f3eb7a8ce93076c05092ba9b58ea0e3e879c65bcfa227a21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://76cff8a2d652df38b04eefc053832baf4cbb5db778ff5e525b18c2825141b118\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://206f6f19fe3197f5d6798cc86c5473bf2decb02a2860ee7ffb5efb5f60952e8c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bafaa9670b05671fb57b658c2960bf3ee32c84e52f86a3a3bbab01d5a38ff0c4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://773a0bc2aab73ffc2717e987543defc9b786cc22300aa54b4ba4f3454504e331\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-06T03:34:22Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1206 03:34:15.305330 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1206 03:34:15.307249 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1592846546/tls.crt::/tmp/serving-cert-1592846546/tls.key\\\\\\\"\\\\nI1206 03:34:21.814037 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1206 03:34:21.816756 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1206 03:34:21.816780 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1206 03:34:21.816806 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1206 03:34:21.816813 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1206 03:34:21.831301 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1206 03:34:21.831311 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1206 03:34:21.831324 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 03:34:21.831330 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 03:34:21.831335 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1206 03:34:21.831339 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1206 03:34:21.831342 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1206 03:34:21.831346 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1206 03:34:21.911269 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:04Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d40e0e7446f8416bf9da94fc53c01be937f9a51a11eb101157d2ddd4597720d2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:04Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2cba89f2f1bed96d59f183628d900d9b72c91eb472e273c8bed0babcab485910\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2cba89f2f1bed96d59f183628d900d9b72c91eb472e273c8bed0babcab485910\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:01Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:54Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:54 crc kubenswrapper[4980]: I1206 03:34:54.067762 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 03:34:54 crc kubenswrapper[4980]: I1206 03:34:54.067835 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 03:34:54 crc kubenswrapper[4980]: E1206 03:34:54.067940 4980 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 06 03:34:54 crc kubenswrapper[4980]: E1206 03:34:54.067967 4980 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 06 03:34:54 crc kubenswrapper[4980]: E1206 03:34:54.068023 4980 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-06 03:35:26.067999995 +0000 UTC m=+85.312751266 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 06 03:34:54 crc kubenswrapper[4980]: E1206 03:34:54.068040 4980 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-06 03:35:26.068031496 +0000 UTC m=+85.312782767 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 06 03:34:54 crc kubenswrapper[4980]: I1206 03:34:54.070881 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://334ea6a273fd7d268906c9f4a6dbf6f4df4fcdbf38f3d7d248cb7d6cf8537794\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ec44dcd5e4bf0a0466bb586ebd61259364dbbfe22499f4cbff854e3bd153082\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:54Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:54 crc kubenswrapper[4980]: I1206 03:34:54.082957 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:22Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:54Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:54 crc kubenswrapper[4980]: I1206 03:34:54.086793 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:34:54 crc kubenswrapper[4980]: I1206 03:34:54.086839 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:34:54 crc kubenswrapper[4980]: I1206 03:34:54.086851 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:34:54 crc kubenswrapper[4980]: I1206 03:34:54.086868 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:34:54 crc kubenswrapper[4980]: I1206 03:34:54.086879 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:34:54Z","lastTransitionTime":"2025-12-06T03:34:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:34:54 crc kubenswrapper[4980]: I1206 03:34:54.095852 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-sv5sz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"873977aa-4eb0-4526-bc0b-757ad2dea2a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01f0cc855c822a71ea8ef6a7ce6e25747ff9956691735973f935c4a28672a7a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lg7tx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:23Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-sv5sz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:54Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:54 crc kubenswrapper[4980]: I1206 03:34:54.116559 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5w4fr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5af0e768-3c9c-4401-ab56-57516cd3170e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:23Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:23Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://768b8e1fe1a00a8467d744da5ad3f61006b0a3612c96f8e4d0ee6fcb93195f9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfzrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6342903b318317ee8d84c2f5d467cb25c811a789be0d6ab0e0cba37b5611e16d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfzrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d620685574db7594132ae24a91ec0a93dce20a9a289540deebb39650b8a1816\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfzrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c67ec8f9c1764fcdf215d52822c2d4f0437dc7f056f05d708767121f6c1f3515\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfzrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f35beae5339eeaf26494ed678e3c1bd02349eae2a4b4615325b17923f3b653b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfzrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a940ec6ee523313728f7e8ae934b7e0b0cff82771e3e64723309db1fae013c3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfzrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fae428d1a6eaeb2e6ddd4a5b1a1170e230635739ea0b39743e565e47fb7ee6e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6cb5cc9f6f22133cebf095db595e5062d799d7e3c8ca0bf31d3809a296e507cc\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-06T03:34:39Z\\\",\\\"message\\\":\\\"lector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1206 03:34:39.306579 6370 reflector.go:311] Stopping reflector *v1.AdminPolicyBasedExternalRoute (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/adminpolicybasedroute/v1/apis/informers/externalversions/factory.go:140\\\\nI1206 03:34:39.306665 6370 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1206 03:34:39.306723 6370 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI1206 03:34:39.306893 6370 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1206 03:34:39.307227 6370 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1206 03:34:39.307306 6370 reflector.go:311] Stopping reflector *v1.NetworkAttachmentDefinition (0s) from github.com/k8snetworkplumbingwg/network-attachment-definition-client/pkg/client/informers/externalversions/factory.go:117\\\\nI1206 03:34:39.307466 6370 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:37Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fae428d1a6eaeb2e6ddd4a5b1a1170e230635739ea0b39743e565e47fb7ee6e4\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-06T03:34:53Z\\\",\\\"message\\\":\\\"/pkg/client/informers/externalversions/factory.go:141\\\\nI1206 03:34:53.908007 6569 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1206 03:34:53.908049 6569 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1206 03:34:53.908094 6569 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1206 03:34:53.908109 6569 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1206 03:34:53.908160 6569 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1206 03:34:53.908208 6569 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1206 03:34:53.908256 6569 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1206 03:34:53.908286 6569 factory.go:656] Stopping watch factory\\\\nI1206 03:34:53.908308 6569 ovnkube.go:599] Stopped ovnkube\\\\nI1206 03:34:53.908349 6569 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1206 03:34:53.908364 6569 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1206 03:34:53.908372 6569 handler.go:208] Removed *v1.Node event handler 2\\\\nI1206 03:34:53.908385 6569 handler.go:208] Removed *v1.Node event handler 7\\\\nI1206 03:34:53.908397 6569 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1206 03:34:53.908415 6569 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1206 03:34:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfzrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe261f4322b85702181d7473ab2a9a4d2d4b869d586c3091f811b12a81a63aae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfzrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d876df04b5ab0ba07a5cbcafbe4648403c5014c59eb7d40adee2657c3a255473\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d876df04b5ab0ba07a5cbcafbe4648403c5014c59eb7d40adee2657c3a255473\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfzrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:23Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-5w4fr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:54Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:54 crc kubenswrapper[4980]: I1206 03:34:54.130717 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:22Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:54Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:54 crc kubenswrapper[4980]: I1206 03:34:54.144535 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-dx44q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5bac22a-512c-49d0-90db-47de6aaaefd0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://97469dea2cd5393da0acf784ccc1969c8a4d07da7d51b14cdc955013ee9710ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f7mdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3b5dab476c7f227a6e4a2eae8469432667fb9fad345d9419ba59feab9fd385e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c3b5dab476c7f227a6e4a2eae8469432667fb9fad345d9419ba59feab9fd385e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f7mdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e921499231cbdf26b8918c98f9ad8c322fd7848865caeef048470de3e2c0efba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e921499231cbdf26b8918c98f9ad8c322fd7848865caeef048470de3e2c0efba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f7mdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7cbfe7d8cdb89916c238296592fab45b9e92e2c585baa6d67230decd1c84c5bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7cbfe7d8cdb89916c238296592fab45b9e92e2c585baa6d67230decd1c84c5bb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f7mdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f45cbf23b6b0d208cefc4b493f901bab9d71ccf2f1456507d54fb1f30c4fd48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7f45cbf23b6b0d208cefc4b493f901bab9d71ccf2f1456507d54fb1f30c4fd48\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f7mdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://344ce4b145ff021fcaf163803d8ad8cc6e9565bf7f0992e2653c43db663d8411\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://344ce4b145ff021fcaf163803d8ad8cc6e9565bf7f0992e2653c43db663d8411\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f7mdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://619e56a9937dd6376ab471ba16da30e5d77f443f5bd33610bea5ec85cd0446ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://619e56a9937dd6376ab471ba16da30e5d77f443f5bd33610bea5ec85cd0446ef\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f7mdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:23Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-dx44q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:54Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:54 crc kubenswrapper[4980]: I1206 03:34:54.155196 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-r5zfb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"320f44d1-a671-4a91-b328-a8b0fdd8f23a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://92393ecaaef8cfcfd01837671f8cc89803dd570b637b4a373b40b7f487788826\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zhm5t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52e8db644463794041978ced2c0591c13e0ada8e5f1a78a94a0a665218a37b5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zhm5t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:23Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-r5zfb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:54Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:54 crc kubenswrapper[4980]: I1206 03:34:54.167737 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-cvzg8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f1cde706-268d-4d26-818f-2f42bfb37b5f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://86f9a77acae360740be827eb0d1e1516e8a49cb9b197b6f5f9468ac70a101158\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rk86l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:29Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-cvzg8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:54Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:54 crc kubenswrapper[4980]: I1206 03:34:54.183555 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 03:34:54 crc kubenswrapper[4980]: I1206 03:34:54.183645 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-mcprh" Dec 06 03:34:54 crc kubenswrapper[4980]: I1206 03:34:54.183555 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 03:34:54 crc kubenswrapper[4980]: E1206 03:34:54.183721 4980 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 03:34:54 crc kubenswrapper[4980]: E1206 03:34:54.183827 4980 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-mcprh" podUID="e6e88c5e-a255-449a-ae38-ac4d73a8e984" Dec 06 03:34:54 crc kubenswrapper[4980]: E1206 03:34:54.183941 4980 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 03:34:54 crc kubenswrapper[4980]: I1206 03:34:54.189746 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cab608b3-a02f-475b-9aae-9b8f5d878de9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f7fcc44b42bcbb55509048e3a566bb557333c4a83e33d7b0fba8eb514317909\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c22e7dd8886c2dfbb742900ed86ec7b0c2edf9a1363e74f6666961d928f06c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://473e792b2d0a9644854ee0581c6120fc9c989bc342899ef72b1d056080fd4694\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ee8e936dfc39361d6f99460910591ca8c901a20f0d274bbb6aaebfa8a00ee06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://77bc12768b018625ba11a6bfd66688f926c4b7f3bc1d0e41c6551a18b485ecf3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://27cf07cd96893900da4afecd0d56fd9269bc60c1d23246b4685fcd8756357c19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://27cf07cd96893900da4afecd0d56fd9269bc60c1d23246b4685fcd8756357c19\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c49feb24566a8284378670fa0390710d38c237acc94ffb654f90f3fa4626e4c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c49feb24566a8284378670fa0390710d38c237acc94ffb654f90f3fa4626e4c8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:03Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://bb74d1c3d752a379fa394c450217badb45511fbcafad62f7b618131cab94b29e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bb74d1c3d752a379fa394c450217badb45511fbcafad62f7b618131cab94b29e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:01Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:54Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:54 crc kubenswrapper[4980]: I1206 03:34:54.190130 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:34:54 crc kubenswrapper[4980]: I1206 03:34:54.190188 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:34:54 crc kubenswrapper[4980]: I1206 03:34:54.190205 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:34:54 crc kubenswrapper[4980]: I1206 03:34:54.190228 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:34:54 crc kubenswrapper[4980]: I1206 03:34:54.190244 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:34:54Z","lastTransitionTime":"2025-12-06T03:34:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:34:54 crc kubenswrapper[4980]: I1206 03:34:54.202344 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://502c727d6baa31f400bee14c34b19abbe38af873f28f6a82a43547431755d981\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:54Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:54 crc kubenswrapper[4980]: I1206 03:34:54.215422 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-kq4z8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f96c1f6c-55e8-4d86-bff3-5381581b7022\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://799498681afa7653408bc23b1e82d75c16e3c809d2ee7f322780e20c537eb710\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bdx7g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:23Z\\\"}}\" for pod \"openshift-multus\"/\"multus-kq4z8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:54Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:54 crc kubenswrapper[4980]: I1206 03:34:54.227776 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-6kxfq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"951b6deb-bcea-4332-8ae8-c23ac7aaef36\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://423d30c3e65b04bdf6ff5ef53d8a7575a434c05e28489370a7370381ebc29de9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b8htf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://282df50dd9653c1cbe97fc2041d9e9ea91a992661682652aafc12e566d8f1b81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b8htf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:37Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-6kxfq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:54Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:54 crc kubenswrapper[4980]: I1206 03:34:54.243332 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-mcprh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e6e88c5e-a255-449a-ae38-ac4d73a8e984\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g7xz9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g7xz9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:38Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-mcprh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:54Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:54 crc kubenswrapper[4980]: I1206 03:34:54.256164 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:54Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:54 crc kubenswrapper[4980]: I1206 03:34:54.270016 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 03:34:54 crc kubenswrapper[4980]: I1206 03:34:54.270177 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 03:34:54 crc kubenswrapper[4980]: I1206 03:34:54.270259 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 03:34:54 crc kubenswrapper[4980]: E1206 03:34:54.270499 4980 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 06 03:34:54 crc kubenswrapper[4980]: E1206 03:34:54.270567 4980 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 06 03:34:54 crc kubenswrapper[4980]: E1206 03:34:54.270584 4980 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 06 03:34:54 crc kubenswrapper[4980]: E1206 03:34:54.270686 4980 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-06 03:35:26.270662627 +0000 UTC m=+85.515413898 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 06 03:34:54 crc kubenswrapper[4980]: E1206 03:34:54.271016 4980 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 03:35:26.270971276 +0000 UTC m=+85.515722537 (durationBeforeRetry 32s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 03:34:54 crc kubenswrapper[4980]: E1206 03:34:54.271229 4980 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 06 03:34:54 crc kubenswrapper[4980]: E1206 03:34:54.271290 4980 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 06 03:34:54 crc kubenswrapper[4980]: E1206 03:34:54.271310 4980 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 06 03:34:54 crc kubenswrapper[4980]: E1206 03:34:54.271409 4980 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-06 03:35:26.271382257 +0000 UTC m=+85.516133688 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 06 03:34:54 crc kubenswrapper[4980]: I1206 03:34:54.273403 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4ef01e7d9809fd9cfc46e3a58b07159cdc6f5a3c210d1c9efc49bca27b30858\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:54Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:54 crc kubenswrapper[4980]: I1206 03:34:54.293454 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:34:54 crc kubenswrapper[4980]: I1206 03:34:54.293547 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:34:54 crc kubenswrapper[4980]: I1206 03:34:54.293567 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:34:54 crc kubenswrapper[4980]: I1206 03:34:54.293595 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:34:54 crc kubenswrapper[4980]: I1206 03:34:54.293614 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:34:54Z","lastTransitionTime":"2025-12-06T03:34:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:34:54 crc kubenswrapper[4980]: I1206 03:34:54.395868 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:34:54 crc kubenswrapper[4980]: I1206 03:34:54.395912 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:34:54 crc kubenswrapper[4980]: I1206 03:34:54.395925 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:34:54 crc kubenswrapper[4980]: I1206 03:34:54.395947 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:34:54 crc kubenswrapper[4980]: I1206 03:34:54.395960 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:34:54Z","lastTransitionTime":"2025-12-06T03:34:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:34:54 crc kubenswrapper[4980]: I1206 03:34:54.472245 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/e6e88c5e-a255-449a-ae38-ac4d73a8e984-metrics-certs\") pod \"network-metrics-daemon-mcprh\" (UID: \"e6e88c5e-a255-449a-ae38-ac4d73a8e984\") " pod="openshift-multus/network-metrics-daemon-mcprh" Dec 06 03:34:54 crc kubenswrapper[4980]: E1206 03:34:54.472411 4980 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 06 03:34:54 crc kubenswrapper[4980]: E1206 03:34:54.472484 4980 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/e6e88c5e-a255-449a-ae38-ac4d73a8e984-metrics-certs podName:e6e88c5e-a255-449a-ae38-ac4d73a8e984 nodeName:}" failed. No retries permitted until 2025-12-06 03:35:10.472463696 +0000 UTC m=+69.717214957 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/e6e88c5e-a255-449a-ae38-ac4d73a8e984-metrics-certs") pod "network-metrics-daemon-mcprh" (UID: "e6e88c5e-a255-449a-ae38-ac4d73a8e984") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 06 03:34:54 crc kubenswrapper[4980]: I1206 03:34:54.499118 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:34:54 crc kubenswrapper[4980]: I1206 03:34:54.499164 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:34:54 crc kubenswrapper[4980]: I1206 03:34:54.499178 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:34:54 crc kubenswrapper[4980]: I1206 03:34:54.499208 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:34:54 crc kubenswrapper[4980]: I1206 03:34:54.499227 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:34:54Z","lastTransitionTime":"2025-12-06T03:34:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:34:54 crc kubenswrapper[4980]: I1206 03:34:54.602647 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:34:54 crc kubenswrapper[4980]: I1206 03:34:54.602726 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:34:54 crc kubenswrapper[4980]: I1206 03:34:54.602747 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:34:54 crc kubenswrapper[4980]: I1206 03:34:54.602775 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:34:54 crc kubenswrapper[4980]: I1206 03:34:54.602799 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:34:54Z","lastTransitionTime":"2025-12-06T03:34:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:34:54 crc kubenswrapper[4980]: I1206 03:34:54.706945 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:34:54 crc kubenswrapper[4980]: I1206 03:34:54.707020 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:34:54 crc kubenswrapper[4980]: I1206 03:34:54.707033 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:34:54 crc kubenswrapper[4980]: I1206 03:34:54.707053 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:34:54 crc kubenswrapper[4980]: I1206 03:34:54.707068 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:34:54Z","lastTransitionTime":"2025-12-06T03:34:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:34:54 crc kubenswrapper[4980]: I1206 03:34:54.810680 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:34:54 crc kubenswrapper[4980]: I1206 03:34:54.810793 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:34:54 crc kubenswrapper[4980]: I1206 03:34:54.810818 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:34:54 crc kubenswrapper[4980]: I1206 03:34:54.810851 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:34:54 crc kubenswrapper[4980]: I1206 03:34:54.810868 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:34:54Z","lastTransitionTime":"2025-12-06T03:34:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:34:54 crc kubenswrapper[4980]: I1206 03:34:54.914836 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:34:54 crc kubenswrapper[4980]: I1206 03:34:54.914907 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:34:54 crc kubenswrapper[4980]: I1206 03:34:54.914916 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:34:54 crc kubenswrapper[4980]: I1206 03:34:54.914939 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:34:54 crc kubenswrapper[4980]: I1206 03:34:54.914950 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:34:54Z","lastTransitionTime":"2025-12-06T03:34:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:34:55 crc kubenswrapper[4980]: I1206 03:34:55.018191 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:34:55 crc kubenswrapper[4980]: I1206 03:34:55.018235 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:34:55 crc kubenswrapper[4980]: I1206 03:34:55.018247 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:34:55 crc kubenswrapper[4980]: I1206 03:34:55.018266 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:34:55 crc kubenswrapper[4980]: I1206 03:34:55.018280 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:34:55Z","lastTransitionTime":"2025-12-06T03:34:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:34:55 crc kubenswrapper[4980]: I1206 03:34:55.024475 4980 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-5w4fr_5af0e768-3c9c-4401-ab56-57516cd3170e/ovnkube-controller/2.log" Dec 06 03:34:55 crc kubenswrapper[4980]: I1206 03:34:55.030470 4980 scope.go:117] "RemoveContainer" containerID="fae428d1a6eaeb2e6ddd4a5b1a1170e230635739ea0b39743e565e47fb7ee6e4" Dec 06 03:34:55 crc kubenswrapper[4980]: E1206 03:34:55.030888 4980 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-5w4fr_openshift-ovn-kubernetes(5af0e768-3c9c-4401-ab56-57516cd3170e)\"" pod="openshift-ovn-kubernetes/ovnkube-node-5w4fr" podUID="5af0e768-3c9c-4401-ab56-57516cd3170e" Dec 06 03:34:55 crc kubenswrapper[4980]: I1206 03:34:55.051494 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:22Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:55Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:55 crc kubenswrapper[4980]: I1206 03:34:55.068714 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://334ea6a273fd7d268906c9f4a6dbf6f4df4fcdbf38f3d7d248cb7d6cf8537794\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ec44dcd5e4bf0a0466bb586ebd61259364dbbfe22499f4cbff854e3bd153082\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:55Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:55 crc kubenswrapper[4980]: I1206 03:34:55.084239 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:22Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:55Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:55 crc kubenswrapper[4980]: I1206 03:34:55.098438 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-sv5sz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"873977aa-4eb0-4526-bc0b-757ad2dea2a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01f0cc855c822a71ea8ef6a7ce6e25747ff9956691735973f935c4a28672a7a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lg7tx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:23Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-sv5sz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:55Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:55 crc kubenswrapper[4980]: I1206 03:34:55.120430 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5w4fr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5af0e768-3c9c-4401-ab56-57516cd3170e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:23Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:23Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://768b8e1fe1a00a8467d744da5ad3f61006b0a3612c96f8e4d0ee6fcb93195f9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfzrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6342903b318317ee8d84c2f5d467cb25c811a789be0d6ab0e0cba37b5611e16d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfzrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d620685574db7594132ae24a91ec0a93dce20a9a289540deebb39650b8a1816\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfzrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c67ec8f9c1764fcdf215d52822c2d4f0437dc7f056f05d708767121f6c1f3515\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfzrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f35beae5339eeaf26494ed678e3c1bd02349eae2a4b4615325b17923f3b653b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfzrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a940ec6ee523313728f7e8ae934b7e0b0cff82771e3e64723309db1fae013c3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfzrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fae428d1a6eaeb2e6ddd4a5b1a1170e230635739ea0b39743e565e47fb7ee6e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fae428d1a6eaeb2e6ddd4a5b1a1170e230635739ea0b39743e565e47fb7ee6e4\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-06T03:34:53Z\\\",\\\"message\\\":\\\"/pkg/client/informers/externalversions/factory.go:141\\\\nI1206 03:34:53.908007 6569 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1206 03:34:53.908049 6569 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1206 03:34:53.908094 6569 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1206 03:34:53.908109 6569 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1206 03:34:53.908160 6569 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1206 03:34:53.908208 6569 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1206 03:34:53.908256 6569 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1206 03:34:53.908286 6569 factory.go:656] Stopping watch factory\\\\nI1206 03:34:53.908308 6569 ovnkube.go:599] Stopped ovnkube\\\\nI1206 03:34:53.908349 6569 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1206 03:34:53.908364 6569 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1206 03:34:53.908372 6569 handler.go:208] Removed *v1.Node event handler 2\\\\nI1206 03:34:53.908385 6569 handler.go:208] Removed *v1.Node event handler 7\\\\nI1206 03:34:53.908397 6569 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1206 03:34:53.908415 6569 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1206 03:34:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:52Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-5w4fr_openshift-ovn-kubernetes(5af0e768-3c9c-4401-ab56-57516cd3170e)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfzrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe261f4322b85702181d7473ab2a9a4d2d4b869d586c3091f811b12a81a63aae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfzrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d876df04b5ab0ba07a5cbcafbe4648403c5014c59eb7d40adee2657c3a255473\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d876df04b5ab0ba07a5cbcafbe4648403c5014c59eb7d40adee2657c3a255473\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfzrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:23Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-5w4fr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:55Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:55 crc kubenswrapper[4980]: I1206 03:34:55.121337 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:34:55 crc kubenswrapper[4980]: I1206 03:34:55.121386 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:34:55 crc kubenswrapper[4980]: I1206 03:34:55.121398 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:34:55 crc kubenswrapper[4980]: I1206 03:34:55.121419 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:34:55 crc kubenswrapper[4980]: I1206 03:34:55.121433 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:34:55Z","lastTransitionTime":"2025-12-06T03:34:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:34:55 crc kubenswrapper[4980]: I1206 03:34:55.134206 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://502c727d6baa31f400bee14c34b19abbe38af873f28f6a82a43547431755d981\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:55Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:55 crc kubenswrapper[4980]: I1206 03:34:55.151298 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-dx44q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5bac22a-512c-49d0-90db-47de6aaaefd0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://97469dea2cd5393da0acf784ccc1969c8a4d07da7d51b14cdc955013ee9710ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f7mdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3b5dab476c7f227a6e4a2eae8469432667fb9fad345d9419ba59feab9fd385e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c3b5dab476c7f227a6e4a2eae8469432667fb9fad345d9419ba59feab9fd385e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f7mdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e921499231cbdf26b8918c98f9ad8c322fd7848865caeef048470de3e2c0efba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e921499231cbdf26b8918c98f9ad8c322fd7848865caeef048470de3e2c0efba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f7mdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7cbfe7d8cdb89916c238296592fab45b9e92e2c585baa6d67230decd1c84c5bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7cbfe7d8cdb89916c238296592fab45b9e92e2c585baa6d67230decd1c84c5bb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f7mdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f45cbf23b6b0d208cefc4b493f901bab9d71ccf2f1456507d54fb1f30c4fd48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7f45cbf23b6b0d208cefc4b493f901bab9d71ccf2f1456507d54fb1f30c4fd48\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f7mdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://344ce4b145ff021fcaf163803d8ad8cc6e9565bf7f0992e2653c43db663d8411\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://344ce4b145ff021fcaf163803d8ad8cc6e9565bf7f0992e2653c43db663d8411\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f7mdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://619e56a9937dd6376ab471ba16da30e5d77f443f5bd33610bea5ec85cd0446ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://619e56a9937dd6376ab471ba16da30e5d77f443f5bd33610bea5ec85cd0446ef\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f7mdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:23Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-dx44q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:55Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:55 crc kubenswrapper[4980]: I1206 03:34:55.165301 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-r5zfb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"320f44d1-a671-4a91-b328-a8b0fdd8f23a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://92393ecaaef8cfcfd01837671f8cc89803dd570b637b4a373b40b7f487788826\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zhm5t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52e8db644463794041978ced2c0591c13e0ada8e5f1a78a94a0a665218a37b5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zhm5t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:23Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-r5zfb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:55Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:55 crc kubenswrapper[4980]: I1206 03:34:55.178430 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-cvzg8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f1cde706-268d-4d26-818f-2f42bfb37b5f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://86f9a77acae360740be827eb0d1e1516e8a49cb9b197b6f5f9468ac70a101158\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rk86l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:29Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-cvzg8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:55Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:55 crc kubenswrapper[4980]: I1206 03:34:55.183678 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 03:34:55 crc kubenswrapper[4980]: E1206 03:34:55.183863 4980 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 03:34:55 crc kubenswrapper[4980]: I1206 03:34:55.201794 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cab608b3-a02f-475b-9aae-9b8f5d878de9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f7fcc44b42bcbb55509048e3a566bb557333c4a83e33d7b0fba8eb514317909\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c22e7dd8886c2dfbb742900ed86ec7b0c2edf9a1363e74f6666961d928f06c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://473e792b2d0a9644854ee0581c6120fc9c989bc342899ef72b1d056080fd4694\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ee8e936dfc39361d6f99460910591ca8c901a20f0d274bbb6aaebfa8a00ee06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://77bc12768b018625ba11a6bfd66688f926c4b7f3bc1d0e41c6551a18b485ecf3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://27cf07cd96893900da4afecd0d56fd9269bc60c1d23246b4685fcd8756357c19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://27cf07cd96893900da4afecd0d56fd9269bc60c1d23246b4685fcd8756357c19\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c49feb24566a8284378670fa0390710d38c237acc94ffb654f90f3fa4626e4c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c49feb24566a8284378670fa0390710d38c237acc94ffb654f90f3fa4626e4c8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:03Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://bb74d1c3d752a379fa394c450217badb45511fbcafad62f7b618131cab94b29e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bb74d1c3d752a379fa394c450217badb45511fbcafad62f7b618131cab94b29e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:01Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:55Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:55 crc kubenswrapper[4980]: I1206 03:34:55.218607 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4ef01e7d9809fd9cfc46e3a58b07159cdc6f5a3c210d1c9efc49bca27b30858\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:55Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:55 crc kubenswrapper[4980]: I1206 03:34:55.223883 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:34:55 crc kubenswrapper[4980]: I1206 03:34:55.223915 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:34:55 crc kubenswrapper[4980]: I1206 03:34:55.223926 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:34:55 crc kubenswrapper[4980]: I1206 03:34:55.223944 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:34:55 crc kubenswrapper[4980]: I1206 03:34:55.223955 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:34:55Z","lastTransitionTime":"2025-12-06T03:34:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:34:55 crc kubenswrapper[4980]: I1206 03:34:55.233832 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-kq4z8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f96c1f6c-55e8-4d86-bff3-5381581b7022\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://799498681afa7653408bc23b1e82d75c16e3c809d2ee7f322780e20c537eb710\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bdx7g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:23Z\\\"}}\" for pod \"openshift-multus\"/\"multus-kq4z8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:55Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:55 crc kubenswrapper[4980]: I1206 03:34:55.282970 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-6kxfq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"951b6deb-bcea-4332-8ae8-c23ac7aaef36\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://423d30c3e65b04bdf6ff5ef53d8a7575a434c05e28489370a7370381ebc29de9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b8htf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://282df50dd9653c1cbe97fc2041d9e9ea91a992661682652aafc12e566d8f1b81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b8htf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:37Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-6kxfq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:55Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:55 crc kubenswrapper[4980]: I1206 03:34:55.295543 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-mcprh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e6e88c5e-a255-449a-ae38-ac4d73a8e984\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g7xz9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g7xz9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:38Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-mcprh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:55Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:55 crc kubenswrapper[4980]: I1206 03:34:55.310339 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:55Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:55 crc kubenswrapper[4980]: I1206 03:34:55.325690 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7f958150-549d-4c80-8a7e-23c39e4f9500\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c28b6895b14e73f4f3eb7a8ce93076c05092ba9b58ea0e3e879c65bcfa227a21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://76cff8a2d652df38b04eefc053832baf4cbb5db778ff5e525b18c2825141b118\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://206f6f19fe3197f5d6798cc86c5473bf2decb02a2860ee7ffb5efb5f60952e8c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bafaa9670b05671fb57b658c2960bf3ee32c84e52f86a3a3bbab01d5a38ff0c4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://773a0bc2aab73ffc2717e987543defc9b786cc22300aa54b4ba4f3454504e331\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-06T03:34:22Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1206 03:34:15.305330 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1206 03:34:15.307249 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1592846546/tls.crt::/tmp/serving-cert-1592846546/tls.key\\\\\\\"\\\\nI1206 03:34:21.814037 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1206 03:34:21.816756 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1206 03:34:21.816780 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1206 03:34:21.816806 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1206 03:34:21.816813 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1206 03:34:21.831301 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1206 03:34:21.831311 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1206 03:34:21.831324 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 03:34:21.831330 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 03:34:21.831335 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1206 03:34:21.831339 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1206 03:34:21.831342 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1206 03:34:21.831346 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1206 03:34:21.911269 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:04Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d40e0e7446f8416bf9da94fc53c01be937f9a51a11eb101157d2ddd4597720d2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:04Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2cba89f2f1bed96d59f183628d900d9b72c91eb472e273c8bed0babcab485910\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2cba89f2f1bed96d59f183628d900d9b72c91eb472e273c8bed0babcab485910\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:01Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:55Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:55 crc kubenswrapper[4980]: I1206 03:34:55.327240 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:34:55 crc kubenswrapper[4980]: I1206 03:34:55.327285 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:34:55 crc kubenswrapper[4980]: I1206 03:34:55.327299 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:34:55 crc kubenswrapper[4980]: I1206 03:34:55.327323 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:34:55 crc kubenswrapper[4980]: I1206 03:34:55.327339 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:34:55Z","lastTransitionTime":"2025-12-06T03:34:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:34:55 crc kubenswrapper[4980]: I1206 03:34:55.338183 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6ed68f66-1cf6-465a-8f66-76d5af6a44d8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://725e84f6bdfb722a8ce69158d03ea4c66a9f087642d55cb9651e003c81be0c8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://71f7b492ea07cdf7dc65fc48184ec1e64914eb1811bbfdddc0fb14c8ad6cb296\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b543aa4ee03d5a14909b67d876c5675c02e1df2a89a3c9ac8d3fb587fea53d8f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://32af3427e872a70cb3c2357116f04cabc8d3d9b69e7306b2e0de2ae22f9469d4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:01Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:55Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:55 crc kubenswrapper[4980]: I1206 03:34:55.429804 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:34:55 crc kubenswrapper[4980]: I1206 03:34:55.429839 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:34:55 crc kubenswrapper[4980]: I1206 03:34:55.429847 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:34:55 crc kubenswrapper[4980]: I1206 03:34:55.429865 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:34:55 crc kubenswrapper[4980]: I1206 03:34:55.429877 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:34:55Z","lastTransitionTime":"2025-12-06T03:34:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:34:55 crc kubenswrapper[4980]: I1206 03:34:55.533900 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:34:55 crc kubenswrapper[4980]: I1206 03:34:55.533960 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:34:55 crc kubenswrapper[4980]: I1206 03:34:55.533978 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:34:55 crc kubenswrapper[4980]: I1206 03:34:55.534001 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:34:55 crc kubenswrapper[4980]: I1206 03:34:55.534014 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:34:55Z","lastTransitionTime":"2025-12-06T03:34:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:34:55 crc kubenswrapper[4980]: I1206 03:34:55.637172 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:34:55 crc kubenswrapper[4980]: I1206 03:34:55.637239 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:34:55 crc kubenswrapper[4980]: I1206 03:34:55.637254 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:34:55 crc kubenswrapper[4980]: I1206 03:34:55.637279 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:34:55 crc kubenswrapper[4980]: I1206 03:34:55.637294 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:34:55Z","lastTransitionTime":"2025-12-06T03:34:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:34:55 crc kubenswrapper[4980]: I1206 03:34:55.740413 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:34:55 crc kubenswrapper[4980]: I1206 03:34:55.740808 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:34:55 crc kubenswrapper[4980]: I1206 03:34:55.740818 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:34:55 crc kubenswrapper[4980]: I1206 03:34:55.740836 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:34:55 crc kubenswrapper[4980]: I1206 03:34:55.740851 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:34:55Z","lastTransitionTime":"2025-12-06T03:34:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:34:55 crc kubenswrapper[4980]: I1206 03:34:55.843353 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:34:55 crc kubenswrapper[4980]: I1206 03:34:55.843388 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:34:55 crc kubenswrapper[4980]: I1206 03:34:55.843397 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:34:55 crc kubenswrapper[4980]: I1206 03:34:55.843414 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:34:55 crc kubenswrapper[4980]: I1206 03:34:55.843425 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:34:55Z","lastTransitionTime":"2025-12-06T03:34:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:34:55 crc kubenswrapper[4980]: I1206 03:34:55.946193 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:34:55 crc kubenswrapper[4980]: I1206 03:34:55.946239 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:34:55 crc kubenswrapper[4980]: I1206 03:34:55.946251 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:34:55 crc kubenswrapper[4980]: I1206 03:34:55.946273 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:34:55 crc kubenswrapper[4980]: I1206 03:34:55.946284 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:34:55Z","lastTransitionTime":"2025-12-06T03:34:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:34:55 crc kubenswrapper[4980]: I1206 03:34:55.995728 4980 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 06 03:34:56 crc kubenswrapper[4980]: I1206 03:34:56.009076 4980 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler/openshift-kube-scheduler-crc"] Dec 06 03:34:56 crc kubenswrapper[4980]: I1206 03:34:56.016312 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:22Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:56Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:56 crc kubenswrapper[4980]: I1206 03:34:56.034004 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://334ea6a273fd7d268906c9f4a6dbf6f4df4fcdbf38f3d7d248cb7d6cf8537794\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ec44dcd5e4bf0a0466bb586ebd61259364dbbfe22499f4cbff854e3bd153082\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:56Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:56 crc kubenswrapper[4980]: I1206 03:34:56.049261 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:34:56 crc kubenswrapper[4980]: I1206 03:34:56.049319 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:34:56 crc kubenswrapper[4980]: I1206 03:34:56.049338 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:34:56 crc kubenswrapper[4980]: I1206 03:34:56.049366 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:34:56 crc kubenswrapper[4980]: I1206 03:34:56.049381 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:34:56Z","lastTransitionTime":"2025-12-06T03:34:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:34:56 crc kubenswrapper[4980]: I1206 03:34:56.049706 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:22Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:56Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:56 crc kubenswrapper[4980]: I1206 03:34:56.063267 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-sv5sz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"873977aa-4eb0-4526-bc0b-757ad2dea2a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01f0cc855c822a71ea8ef6a7ce6e25747ff9956691735973f935c4a28672a7a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lg7tx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:23Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-sv5sz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:56Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:56 crc kubenswrapper[4980]: I1206 03:34:56.085044 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5w4fr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5af0e768-3c9c-4401-ab56-57516cd3170e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:23Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:23Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://768b8e1fe1a00a8467d744da5ad3f61006b0a3612c96f8e4d0ee6fcb93195f9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfzrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6342903b318317ee8d84c2f5d467cb25c811a789be0d6ab0e0cba37b5611e16d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfzrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d620685574db7594132ae24a91ec0a93dce20a9a289540deebb39650b8a1816\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfzrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c67ec8f9c1764fcdf215d52822c2d4f0437dc7f056f05d708767121f6c1f3515\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfzrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f35beae5339eeaf26494ed678e3c1bd02349eae2a4b4615325b17923f3b653b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfzrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a940ec6ee523313728f7e8ae934b7e0b0cff82771e3e64723309db1fae013c3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfzrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fae428d1a6eaeb2e6ddd4a5b1a1170e230635739ea0b39743e565e47fb7ee6e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fae428d1a6eaeb2e6ddd4a5b1a1170e230635739ea0b39743e565e47fb7ee6e4\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-06T03:34:53Z\\\",\\\"message\\\":\\\"/pkg/client/informers/externalversions/factory.go:141\\\\nI1206 03:34:53.908007 6569 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1206 03:34:53.908049 6569 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1206 03:34:53.908094 6569 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1206 03:34:53.908109 6569 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1206 03:34:53.908160 6569 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1206 03:34:53.908208 6569 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1206 03:34:53.908256 6569 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1206 03:34:53.908286 6569 factory.go:656] Stopping watch factory\\\\nI1206 03:34:53.908308 6569 ovnkube.go:599] Stopped ovnkube\\\\nI1206 03:34:53.908349 6569 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1206 03:34:53.908364 6569 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1206 03:34:53.908372 6569 handler.go:208] Removed *v1.Node event handler 2\\\\nI1206 03:34:53.908385 6569 handler.go:208] Removed *v1.Node event handler 7\\\\nI1206 03:34:53.908397 6569 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1206 03:34:53.908415 6569 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1206 03:34:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:52Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-5w4fr_openshift-ovn-kubernetes(5af0e768-3c9c-4401-ab56-57516cd3170e)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfzrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe261f4322b85702181d7473ab2a9a4d2d4b869d586c3091f811b12a81a63aae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfzrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d876df04b5ab0ba07a5cbcafbe4648403c5014c59eb7d40adee2657c3a255473\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d876df04b5ab0ba07a5cbcafbe4648403c5014c59eb7d40adee2657c3a255473\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfzrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:23Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-5w4fr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:56Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:56 crc kubenswrapper[4980]: I1206 03:34:56.113846 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cab608b3-a02f-475b-9aae-9b8f5d878de9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f7fcc44b42bcbb55509048e3a566bb557333c4a83e33d7b0fba8eb514317909\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c22e7dd8886c2dfbb742900ed86ec7b0c2edf9a1363e74f6666961d928f06c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://473e792b2d0a9644854ee0581c6120fc9c989bc342899ef72b1d056080fd4694\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ee8e936dfc39361d6f99460910591ca8c901a20f0d274bbb6aaebfa8a00ee06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://77bc12768b018625ba11a6bfd66688f926c4b7f3bc1d0e41c6551a18b485ecf3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://27cf07cd96893900da4afecd0d56fd9269bc60c1d23246b4685fcd8756357c19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://27cf07cd96893900da4afecd0d56fd9269bc60c1d23246b4685fcd8756357c19\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c49feb24566a8284378670fa0390710d38c237acc94ffb654f90f3fa4626e4c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c49feb24566a8284378670fa0390710d38c237acc94ffb654f90f3fa4626e4c8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:03Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://bb74d1c3d752a379fa394c450217badb45511fbcafad62f7b618131cab94b29e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bb74d1c3d752a379fa394c450217badb45511fbcafad62f7b618131cab94b29e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:01Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:56Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:56 crc kubenswrapper[4980]: I1206 03:34:56.129014 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://502c727d6baa31f400bee14c34b19abbe38af873f28f6a82a43547431755d981\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:56Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:56 crc kubenswrapper[4980]: I1206 03:34:56.147174 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-dx44q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5bac22a-512c-49d0-90db-47de6aaaefd0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://97469dea2cd5393da0acf784ccc1969c8a4d07da7d51b14cdc955013ee9710ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f7mdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3b5dab476c7f227a6e4a2eae8469432667fb9fad345d9419ba59feab9fd385e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c3b5dab476c7f227a6e4a2eae8469432667fb9fad345d9419ba59feab9fd385e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f7mdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e921499231cbdf26b8918c98f9ad8c322fd7848865caeef048470de3e2c0efba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e921499231cbdf26b8918c98f9ad8c322fd7848865caeef048470de3e2c0efba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f7mdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7cbfe7d8cdb89916c238296592fab45b9e92e2c585baa6d67230decd1c84c5bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7cbfe7d8cdb89916c238296592fab45b9e92e2c585baa6d67230decd1c84c5bb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f7mdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f45cbf23b6b0d208cefc4b493f901bab9d71ccf2f1456507d54fb1f30c4fd48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7f45cbf23b6b0d208cefc4b493f901bab9d71ccf2f1456507d54fb1f30c4fd48\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f7mdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://344ce4b145ff021fcaf163803d8ad8cc6e9565bf7f0992e2653c43db663d8411\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://344ce4b145ff021fcaf163803d8ad8cc6e9565bf7f0992e2653c43db663d8411\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f7mdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://619e56a9937dd6376ab471ba16da30e5d77f443f5bd33610bea5ec85cd0446ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://619e56a9937dd6376ab471ba16da30e5d77f443f5bd33610bea5ec85cd0446ef\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f7mdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:23Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-dx44q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:56Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:56 crc kubenswrapper[4980]: I1206 03:34:56.151982 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:34:56 crc kubenswrapper[4980]: I1206 03:34:56.152065 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:34:56 crc kubenswrapper[4980]: I1206 03:34:56.152077 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:34:56 crc kubenswrapper[4980]: I1206 03:34:56.152096 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:34:56 crc kubenswrapper[4980]: I1206 03:34:56.152109 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:34:56Z","lastTransitionTime":"2025-12-06T03:34:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:34:56 crc kubenswrapper[4980]: I1206 03:34:56.168456 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-r5zfb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"320f44d1-a671-4a91-b328-a8b0fdd8f23a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://92393ecaaef8cfcfd01837671f8cc89803dd570b637b4a373b40b7f487788826\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zhm5t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52e8db644463794041978ced2c0591c13e0ada8e5f1a78a94a0a665218a37b5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zhm5t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:23Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-r5zfb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:56Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:56 crc kubenswrapper[4980]: I1206 03:34:56.183752 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 03:34:56 crc kubenswrapper[4980]: I1206 03:34:56.183814 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-mcprh" Dec 06 03:34:56 crc kubenswrapper[4980]: I1206 03:34:56.183849 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 03:34:56 crc kubenswrapper[4980]: E1206 03:34:56.183973 4980 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 03:34:56 crc kubenswrapper[4980]: E1206 03:34:56.184100 4980 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 03:34:56 crc kubenswrapper[4980]: E1206 03:34:56.184314 4980 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-mcprh" podUID="e6e88c5e-a255-449a-ae38-ac4d73a8e984" Dec 06 03:34:56 crc kubenswrapper[4980]: I1206 03:34:56.193663 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-cvzg8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f1cde706-268d-4d26-818f-2f42bfb37b5f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://86f9a77acae360740be827eb0d1e1516e8a49cb9b197b6f5f9468ac70a101158\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rk86l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:29Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-cvzg8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:56Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:56 crc kubenswrapper[4980]: I1206 03:34:56.215027 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:56Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:56 crc kubenswrapper[4980]: I1206 03:34:56.237405 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4ef01e7d9809fd9cfc46e3a58b07159cdc6f5a3c210d1c9efc49bca27b30858\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:56Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:56 crc kubenswrapper[4980]: I1206 03:34:56.253311 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-kq4z8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f96c1f6c-55e8-4d86-bff3-5381581b7022\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://799498681afa7653408bc23b1e82d75c16e3c809d2ee7f322780e20c537eb710\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bdx7g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:23Z\\\"}}\" for pod \"openshift-multus\"/\"multus-kq4z8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:56Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:56 crc kubenswrapper[4980]: I1206 03:34:56.255486 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:34:56 crc kubenswrapper[4980]: I1206 03:34:56.255567 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:34:56 crc kubenswrapper[4980]: I1206 03:34:56.255579 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:34:56 crc kubenswrapper[4980]: I1206 03:34:56.255608 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:34:56 crc kubenswrapper[4980]: I1206 03:34:56.255630 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:34:56Z","lastTransitionTime":"2025-12-06T03:34:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:34:56 crc kubenswrapper[4980]: I1206 03:34:56.266758 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-6kxfq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"951b6deb-bcea-4332-8ae8-c23ac7aaef36\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://423d30c3e65b04bdf6ff5ef53d8a7575a434c05e28489370a7370381ebc29de9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b8htf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://282df50dd9653c1cbe97fc2041d9e9ea91a992661682652aafc12e566d8f1b81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b8htf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:37Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-6kxfq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:56Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:56 crc kubenswrapper[4980]: I1206 03:34:56.278270 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-mcprh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e6e88c5e-a255-449a-ae38-ac4d73a8e984\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g7xz9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g7xz9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:38Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-mcprh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:56Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:56 crc kubenswrapper[4980]: I1206 03:34:56.294953 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7f958150-549d-4c80-8a7e-23c39e4f9500\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c28b6895b14e73f4f3eb7a8ce93076c05092ba9b58ea0e3e879c65bcfa227a21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://76cff8a2d652df38b04eefc053832baf4cbb5db778ff5e525b18c2825141b118\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://206f6f19fe3197f5d6798cc86c5473bf2decb02a2860ee7ffb5efb5f60952e8c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bafaa9670b05671fb57b658c2960bf3ee32c84e52f86a3a3bbab01d5a38ff0c4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://773a0bc2aab73ffc2717e987543defc9b786cc22300aa54b4ba4f3454504e331\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-06T03:34:22Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1206 03:34:15.305330 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1206 03:34:15.307249 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1592846546/tls.crt::/tmp/serving-cert-1592846546/tls.key\\\\\\\"\\\\nI1206 03:34:21.814037 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1206 03:34:21.816756 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1206 03:34:21.816780 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1206 03:34:21.816806 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1206 03:34:21.816813 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1206 03:34:21.831301 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1206 03:34:21.831311 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1206 03:34:21.831324 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 03:34:21.831330 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 03:34:21.831335 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1206 03:34:21.831339 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1206 03:34:21.831342 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1206 03:34:21.831346 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1206 03:34:21.911269 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:04Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d40e0e7446f8416bf9da94fc53c01be937f9a51a11eb101157d2ddd4597720d2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:04Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2cba89f2f1bed96d59f183628d900d9b72c91eb472e273c8bed0babcab485910\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2cba89f2f1bed96d59f183628d900d9b72c91eb472e273c8bed0babcab485910\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:01Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:56Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:56 crc kubenswrapper[4980]: I1206 03:34:56.309741 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6ed68f66-1cf6-465a-8f66-76d5af6a44d8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://725e84f6bdfb722a8ce69158d03ea4c66a9f087642d55cb9651e003c81be0c8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://71f7b492ea07cdf7dc65fc48184ec1e64914eb1811bbfdddc0fb14c8ad6cb296\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b543aa4ee03d5a14909b67d876c5675c02e1df2a89a3c9ac8d3fb587fea53d8f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://32af3427e872a70cb3c2357116f04cabc8d3d9b69e7306b2e0de2ae22f9469d4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:01Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:56Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:56 crc kubenswrapper[4980]: I1206 03:34:56.359326 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:34:56 crc kubenswrapper[4980]: I1206 03:34:56.359361 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:34:56 crc kubenswrapper[4980]: I1206 03:34:56.359370 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:34:56 crc kubenswrapper[4980]: I1206 03:34:56.359389 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:34:56 crc kubenswrapper[4980]: I1206 03:34:56.359401 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:34:56Z","lastTransitionTime":"2025-12-06T03:34:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:34:56 crc kubenswrapper[4980]: I1206 03:34:56.462371 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:34:56 crc kubenswrapper[4980]: I1206 03:34:56.462408 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:34:56 crc kubenswrapper[4980]: I1206 03:34:56.462420 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:34:56 crc kubenswrapper[4980]: I1206 03:34:56.462437 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:34:56 crc kubenswrapper[4980]: I1206 03:34:56.462447 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:34:56Z","lastTransitionTime":"2025-12-06T03:34:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:34:56 crc kubenswrapper[4980]: I1206 03:34:56.565415 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:34:56 crc kubenswrapper[4980]: I1206 03:34:56.565461 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:34:56 crc kubenswrapper[4980]: I1206 03:34:56.565474 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:34:56 crc kubenswrapper[4980]: I1206 03:34:56.565497 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:34:56 crc kubenswrapper[4980]: I1206 03:34:56.565530 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:34:56Z","lastTransitionTime":"2025-12-06T03:34:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:34:56 crc kubenswrapper[4980]: I1206 03:34:56.668845 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:34:56 crc kubenswrapper[4980]: I1206 03:34:56.668902 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:34:56 crc kubenswrapper[4980]: I1206 03:34:56.668913 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:34:56 crc kubenswrapper[4980]: I1206 03:34:56.668933 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:34:56 crc kubenswrapper[4980]: I1206 03:34:56.668944 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:34:56Z","lastTransitionTime":"2025-12-06T03:34:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:34:56 crc kubenswrapper[4980]: I1206 03:34:56.772306 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:34:56 crc kubenswrapper[4980]: I1206 03:34:56.772354 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:34:56 crc kubenswrapper[4980]: I1206 03:34:56.772364 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:34:56 crc kubenswrapper[4980]: I1206 03:34:56.772382 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:34:56 crc kubenswrapper[4980]: I1206 03:34:56.772398 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:34:56Z","lastTransitionTime":"2025-12-06T03:34:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:34:56 crc kubenswrapper[4980]: I1206 03:34:56.874810 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:34:56 crc kubenswrapper[4980]: I1206 03:34:56.874853 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:34:56 crc kubenswrapper[4980]: I1206 03:34:56.874870 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:34:56 crc kubenswrapper[4980]: I1206 03:34:56.874888 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:34:56 crc kubenswrapper[4980]: I1206 03:34:56.874899 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:34:56Z","lastTransitionTime":"2025-12-06T03:34:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:34:56 crc kubenswrapper[4980]: I1206 03:34:56.978406 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:34:56 crc kubenswrapper[4980]: I1206 03:34:56.978460 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:34:56 crc kubenswrapper[4980]: I1206 03:34:56.978476 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:34:56 crc kubenswrapper[4980]: I1206 03:34:56.978496 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:34:56 crc kubenswrapper[4980]: I1206 03:34:56.978532 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:34:56Z","lastTransitionTime":"2025-12-06T03:34:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:34:57 crc kubenswrapper[4980]: I1206 03:34:57.082368 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:34:57 crc kubenswrapper[4980]: I1206 03:34:57.082447 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:34:57 crc kubenswrapper[4980]: I1206 03:34:57.082467 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:34:57 crc kubenswrapper[4980]: I1206 03:34:57.082549 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:34:57 crc kubenswrapper[4980]: I1206 03:34:57.082586 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:34:57Z","lastTransitionTime":"2025-12-06T03:34:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:34:57 crc kubenswrapper[4980]: I1206 03:34:57.183795 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 03:34:57 crc kubenswrapper[4980]: E1206 03:34:57.183956 4980 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 03:34:57 crc kubenswrapper[4980]: I1206 03:34:57.186404 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:34:57 crc kubenswrapper[4980]: I1206 03:34:57.186462 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:34:57 crc kubenswrapper[4980]: I1206 03:34:57.186473 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:34:57 crc kubenswrapper[4980]: I1206 03:34:57.186503 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:34:57 crc kubenswrapper[4980]: I1206 03:34:57.186557 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:34:57Z","lastTransitionTime":"2025-12-06T03:34:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:34:57 crc kubenswrapper[4980]: I1206 03:34:57.289434 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:34:57 crc kubenswrapper[4980]: I1206 03:34:57.289489 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:34:57 crc kubenswrapper[4980]: I1206 03:34:57.289504 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:34:57 crc kubenswrapper[4980]: I1206 03:34:57.289540 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:34:57 crc kubenswrapper[4980]: I1206 03:34:57.289551 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:34:57Z","lastTransitionTime":"2025-12-06T03:34:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:34:57 crc kubenswrapper[4980]: I1206 03:34:57.393667 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:34:57 crc kubenswrapper[4980]: I1206 03:34:57.393722 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:34:57 crc kubenswrapper[4980]: I1206 03:34:57.393741 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:34:57 crc kubenswrapper[4980]: I1206 03:34:57.393762 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:34:57 crc kubenswrapper[4980]: I1206 03:34:57.393779 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:34:57Z","lastTransitionTime":"2025-12-06T03:34:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:34:57 crc kubenswrapper[4980]: I1206 03:34:57.496321 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:34:57 crc kubenswrapper[4980]: I1206 03:34:57.496363 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:34:57 crc kubenswrapper[4980]: I1206 03:34:57.496372 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:34:57 crc kubenswrapper[4980]: I1206 03:34:57.496390 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:34:57 crc kubenswrapper[4980]: I1206 03:34:57.496399 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:34:57Z","lastTransitionTime":"2025-12-06T03:34:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:34:57 crc kubenswrapper[4980]: I1206 03:34:57.599232 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:34:57 crc kubenswrapper[4980]: I1206 03:34:57.599279 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:34:57 crc kubenswrapper[4980]: I1206 03:34:57.599291 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:34:57 crc kubenswrapper[4980]: I1206 03:34:57.599309 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:34:57 crc kubenswrapper[4980]: I1206 03:34:57.599322 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:34:57Z","lastTransitionTime":"2025-12-06T03:34:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:34:57 crc kubenswrapper[4980]: I1206 03:34:57.701639 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:34:57 crc kubenswrapper[4980]: I1206 03:34:57.701686 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:34:57 crc kubenswrapper[4980]: I1206 03:34:57.701695 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:34:57 crc kubenswrapper[4980]: I1206 03:34:57.701713 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:34:57 crc kubenswrapper[4980]: I1206 03:34:57.701725 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:34:57Z","lastTransitionTime":"2025-12-06T03:34:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:34:57 crc kubenswrapper[4980]: I1206 03:34:57.805023 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:34:57 crc kubenswrapper[4980]: I1206 03:34:57.805077 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:34:57 crc kubenswrapper[4980]: I1206 03:34:57.805088 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:34:57 crc kubenswrapper[4980]: I1206 03:34:57.805109 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:34:57 crc kubenswrapper[4980]: I1206 03:34:57.805121 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:34:57Z","lastTransitionTime":"2025-12-06T03:34:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:34:57 crc kubenswrapper[4980]: I1206 03:34:57.912859 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:34:57 crc kubenswrapper[4980]: I1206 03:34:57.912922 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:34:57 crc kubenswrapper[4980]: I1206 03:34:57.912931 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:34:57 crc kubenswrapper[4980]: I1206 03:34:57.912950 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:34:57 crc kubenswrapper[4980]: I1206 03:34:57.912961 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:34:57Z","lastTransitionTime":"2025-12-06T03:34:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:34:58 crc kubenswrapper[4980]: I1206 03:34:58.015309 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:34:58 crc kubenswrapper[4980]: I1206 03:34:58.015368 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:34:58 crc kubenswrapper[4980]: I1206 03:34:58.015381 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:34:58 crc kubenswrapper[4980]: I1206 03:34:58.015400 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:34:58 crc kubenswrapper[4980]: I1206 03:34:58.015418 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:34:58Z","lastTransitionTime":"2025-12-06T03:34:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:34:58 crc kubenswrapper[4980]: I1206 03:34:58.118269 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:34:58 crc kubenswrapper[4980]: I1206 03:34:58.118320 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:34:58 crc kubenswrapper[4980]: I1206 03:34:58.118332 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:34:58 crc kubenswrapper[4980]: I1206 03:34:58.118353 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:34:58 crc kubenswrapper[4980]: I1206 03:34:58.118366 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:34:58Z","lastTransitionTime":"2025-12-06T03:34:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:34:58 crc kubenswrapper[4980]: I1206 03:34:58.183589 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 03:34:58 crc kubenswrapper[4980]: I1206 03:34:58.183716 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 03:34:58 crc kubenswrapper[4980]: I1206 03:34:58.183611 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-mcprh" Dec 06 03:34:58 crc kubenswrapper[4980]: E1206 03:34:58.183929 4980 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 03:34:58 crc kubenswrapper[4980]: E1206 03:34:58.184400 4980 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 03:34:58 crc kubenswrapper[4980]: E1206 03:34:58.184292 4980 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-mcprh" podUID="e6e88c5e-a255-449a-ae38-ac4d73a8e984" Dec 06 03:34:58 crc kubenswrapper[4980]: I1206 03:34:58.220546 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:34:58 crc kubenswrapper[4980]: I1206 03:34:58.220601 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:34:58 crc kubenswrapper[4980]: I1206 03:34:58.220614 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:34:58 crc kubenswrapper[4980]: I1206 03:34:58.220637 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:34:58 crc kubenswrapper[4980]: I1206 03:34:58.220651 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:34:58Z","lastTransitionTime":"2025-12-06T03:34:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:34:58 crc kubenswrapper[4980]: I1206 03:34:58.324180 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:34:58 crc kubenswrapper[4980]: I1206 03:34:58.324227 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:34:58 crc kubenswrapper[4980]: I1206 03:34:58.324236 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:34:58 crc kubenswrapper[4980]: I1206 03:34:58.324252 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:34:58 crc kubenswrapper[4980]: I1206 03:34:58.324261 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:34:58Z","lastTransitionTime":"2025-12-06T03:34:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:34:58 crc kubenswrapper[4980]: I1206 03:34:58.427805 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:34:58 crc kubenswrapper[4980]: I1206 03:34:58.427870 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:34:58 crc kubenswrapper[4980]: I1206 03:34:58.427884 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:34:58 crc kubenswrapper[4980]: I1206 03:34:58.427903 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:34:58 crc kubenswrapper[4980]: I1206 03:34:58.427914 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:34:58Z","lastTransitionTime":"2025-12-06T03:34:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:34:58 crc kubenswrapper[4980]: I1206 03:34:58.531278 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:34:58 crc kubenswrapper[4980]: I1206 03:34:58.531391 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:34:58 crc kubenswrapper[4980]: I1206 03:34:58.531402 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:34:58 crc kubenswrapper[4980]: I1206 03:34:58.531422 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:34:58 crc kubenswrapper[4980]: I1206 03:34:58.531434 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:34:58Z","lastTransitionTime":"2025-12-06T03:34:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:34:58 crc kubenswrapper[4980]: I1206 03:34:58.634866 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:34:58 crc kubenswrapper[4980]: I1206 03:34:58.634936 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:34:58 crc kubenswrapper[4980]: I1206 03:34:58.634949 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:34:58 crc kubenswrapper[4980]: I1206 03:34:58.634971 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:34:58 crc kubenswrapper[4980]: I1206 03:34:58.634987 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:34:58Z","lastTransitionTime":"2025-12-06T03:34:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:34:58 crc kubenswrapper[4980]: I1206 03:34:58.738642 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:34:58 crc kubenswrapper[4980]: I1206 03:34:58.738675 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:34:58 crc kubenswrapper[4980]: I1206 03:34:58.738685 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:34:58 crc kubenswrapper[4980]: I1206 03:34:58.738702 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:34:58 crc kubenswrapper[4980]: I1206 03:34:58.738712 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:34:58Z","lastTransitionTime":"2025-12-06T03:34:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:34:58 crc kubenswrapper[4980]: I1206 03:34:58.842124 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:34:58 crc kubenswrapper[4980]: I1206 03:34:58.842160 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:34:58 crc kubenswrapper[4980]: I1206 03:34:58.842169 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:34:58 crc kubenswrapper[4980]: I1206 03:34:58.842185 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:34:58 crc kubenswrapper[4980]: I1206 03:34:58.842196 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:34:58Z","lastTransitionTime":"2025-12-06T03:34:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:34:58 crc kubenswrapper[4980]: I1206 03:34:58.946153 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:34:58 crc kubenswrapper[4980]: I1206 03:34:58.946233 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:34:58 crc kubenswrapper[4980]: I1206 03:34:58.946252 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:34:58 crc kubenswrapper[4980]: I1206 03:34:58.946282 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:34:58 crc kubenswrapper[4980]: I1206 03:34:58.946300 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:34:58Z","lastTransitionTime":"2025-12-06T03:34:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:34:59 crc kubenswrapper[4980]: I1206 03:34:59.049805 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:34:59 crc kubenswrapper[4980]: I1206 03:34:59.049889 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:34:59 crc kubenswrapper[4980]: I1206 03:34:59.049906 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:34:59 crc kubenswrapper[4980]: I1206 03:34:59.049933 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:34:59 crc kubenswrapper[4980]: I1206 03:34:59.049950 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:34:59Z","lastTransitionTime":"2025-12-06T03:34:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:34:59 crc kubenswrapper[4980]: I1206 03:34:59.153503 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:34:59 crc kubenswrapper[4980]: I1206 03:34:59.153655 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:34:59 crc kubenswrapper[4980]: I1206 03:34:59.153679 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:34:59 crc kubenswrapper[4980]: I1206 03:34:59.153708 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:34:59 crc kubenswrapper[4980]: I1206 03:34:59.153732 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:34:59Z","lastTransitionTime":"2025-12-06T03:34:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:34:59 crc kubenswrapper[4980]: I1206 03:34:59.155668 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:34:59 crc kubenswrapper[4980]: I1206 03:34:59.155742 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:34:59 crc kubenswrapper[4980]: I1206 03:34:59.155765 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:34:59 crc kubenswrapper[4980]: I1206 03:34:59.155797 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:34:59 crc kubenswrapper[4980]: I1206 03:34:59.155824 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:34:59Z","lastTransitionTime":"2025-12-06T03:34:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:34:59 crc kubenswrapper[4980]: E1206 03:34:59.176322 4980 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T03:34:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:59Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T03:34:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:59Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T03:34:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:59Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T03:34:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:59Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"9b4ed4f8-ed8d-447b-9e1f-13defbdcf0f7\\\",\\\"systemUUID\\\":\\\"a5b493aa-69de-4688-8ff6-1e5f49b7f014\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:59Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:59 crc kubenswrapper[4980]: I1206 03:34:59.182470 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:34:59 crc kubenswrapper[4980]: I1206 03:34:59.182584 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:34:59 crc kubenswrapper[4980]: I1206 03:34:59.182610 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:34:59 crc kubenswrapper[4980]: I1206 03:34:59.182640 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:34:59 crc kubenswrapper[4980]: I1206 03:34:59.182663 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:34:59Z","lastTransitionTime":"2025-12-06T03:34:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:34:59 crc kubenswrapper[4980]: I1206 03:34:59.183692 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 03:34:59 crc kubenswrapper[4980]: E1206 03:34:59.183910 4980 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 03:34:59 crc kubenswrapper[4980]: E1206 03:34:59.202186 4980 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T03:34:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:59Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T03:34:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:59Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T03:34:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:59Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T03:34:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:59Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"9b4ed4f8-ed8d-447b-9e1f-13defbdcf0f7\\\",\\\"systemUUID\\\":\\\"a5b493aa-69de-4688-8ff6-1e5f49b7f014\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:59Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:59 crc kubenswrapper[4980]: I1206 03:34:59.207462 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:34:59 crc kubenswrapper[4980]: I1206 03:34:59.207744 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:34:59 crc kubenswrapper[4980]: I1206 03:34:59.207948 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:34:59 crc kubenswrapper[4980]: I1206 03:34:59.208168 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:34:59 crc kubenswrapper[4980]: I1206 03:34:59.208369 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:34:59Z","lastTransitionTime":"2025-12-06T03:34:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:34:59 crc kubenswrapper[4980]: E1206 03:34:59.226410 4980 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T03:34:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:59Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T03:34:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:59Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T03:34:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:59Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T03:34:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:59Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"9b4ed4f8-ed8d-447b-9e1f-13defbdcf0f7\\\",\\\"systemUUID\\\":\\\"a5b493aa-69de-4688-8ff6-1e5f49b7f014\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:59Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:59 crc kubenswrapper[4980]: I1206 03:34:59.232177 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:34:59 crc kubenswrapper[4980]: I1206 03:34:59.232244 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:34:59 crc kubenswrapper[4980]: I1206 03:34:59.232269 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:34:59 crc kubenswrapper[4980]: I1206 03:34:59.232300 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:34:59 crc kubenswrapper[4980]: I1206 03:34:59.232324 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:34:59Z","lastTransitionTime":"2025-12-06T03:34:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:34:59 crc kubenswrapper[4980]: E1206 03:34:59.251320 4980 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T03:34:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:59Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T03:34:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:59Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T03:34:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:59Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T03:34:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:59Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"9b4ed4f8-ed8d-447b-9e1f-13defbdcf0f7\\\",\\\"systemUUID\\\":\\\"a5b493aa-69de-4688-8ff6-1e5f49b7f014\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:59Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:59 crc kubenswrapper[4980]: I1206 03:34:59.256683 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:34:59 crc kubenswrapper[4980]: I1206 03:34:59.256723 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:34:59 crc kubenswrapper[4980]: I1206 03:34:59.256740 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:34:59 crc kubenswrapper[4980]: I1206 03:34:59.256763 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:34:59 crc kubenswrapper[4980]: I1206 03:34:59.256778 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:34:59Z","lastTransitionTime":"2025-12-06T03:34:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:34:59 crc kubenswrapper[4980]: E1206 03:34:59.272429 4980 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T03:34:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:59Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T03:34:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:59Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T03:34:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:59Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T03:34:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:59Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"9b4ed4f8-ed8d-447b-9e1f-13defbdcf0f7\\\",\\\"systemUUID\\\":\\\"a5b493aa-69de-4688-8ff6-1e5f49b7f014\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:34:59Z is after 2025-08-24T17:21:41Z" Dec 06 03:34:59 crc kubenswrapper[4980]: E1206 03:34:59.272697 4980 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 06 03:34:59 crc kubenswrapper[4980]: I1206 03:34:59.274458 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:34:59 crc kubenswrapper[4980]: I1206 03:34:59.274783 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:34:59 crc kubenswrapper[4980]: I1206 03:34:59.274843 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:34:59 crc kubenswrapper[4980]: I1206 03:34:59.274869 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:34:59 crc kubenswrapper[4980]: I1206 03:34:59.275184 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:34:59Z","lastTransitionTime":"2025-12-06T03:34:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:34:59 crc kubenswrapper[4980]: I1206 03:34:59.377859 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:34:59 crc kubenswrapper[4980]: I1206 03:34:59.378173 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:34:59 crc kubenswrapper[4980]: I1206 03:34:59.378271 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:34:59 crc kubenswrapper[4980]: I1206 03:34:59.378370 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:34:59 crc kubenswrapper[4980]: I1206 03:34:59.378499 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:34:59Z","lastTransitionTime":"2025-12-06T03:34:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:34:59 crc kubenswrapper[4980]: I1206 03:34:59.480976 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:34:59 crc kubenswrapper[4980]: I1206 03:34:59.481285 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:34:59 crc kubenswrapper[4980]: I1206 03:34:59.481411 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:34:59 crc kubenswrapper[4980]: I1206 03:34:59.481555 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:34:59 crc kubenswrapper[4980]: I1206 03:34:59.481669 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:34:59Z","lastTransitionTime":"2025-12-06T03:34:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:34:59 crc kubenswrapper[4980]: I1206 03:34:59.584369 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:34:59 crc kubenswrapper[4980]: I1206 03:34:59.584438 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:34:59 crc kubenswrapper[4980]: I1206 03:34:59.584457 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:34:59 crc kubenswrapper[4980]: I1206 03:34:59.584482 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:34:59 crc kubenswrapper[4980]: I1206 03:34:59.584500 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:34:59Z","lastTransitionTime":"2025-12-06T03:34:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:34:59 crc kubenswrapper[4980]: I1206 03:34:59.687599 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:34:59 crc kubenswrapper[4980]: I1206 03:34:59.688031 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:34:59 crc kubenswrapper[4980]: I1206 03:34:59.688211 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:34:59 crc kubenswrapper[4980]: I1206 03:34:59.688362 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:34:59 crc kubenswrapper[4980]: I1206 03:34:59.688577 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:34:59Z","lastTransitionTime":"2025-12-06T03:34:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:34:59 crc kubenswrapper[4980]: I1206 03:34:59.792256 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:34:59 crc kubenswrapper[4980]: I1206 03:34:59.792308 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:34:59 crc kubenswrapper[4980]: I1206 03:34:59.792323 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:34:59 crc kubenswrapper[4980]: I1206 03:34:59.792343 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:34:59 crc kubenswrapper[4980]: I1206 03:34:59.792358 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:34:59Z","lastTransitionTime":"2025-12-06T03:34:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:34:59 crc kubenswrapper[4980]: I1206 03:34:59.896034 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:34:59 crc kubenswrapper[4980]: I1206 03:34:59.896105 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:34:59 crc kubenswrapper[4980]: I1206 03:34:59.896123 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:34:59 crc kubenswrapper[4980]: I1206 03:34:59.896148 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:34:59 crc kubenswrapper[4980]: I1206 03:34:59.896165 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:34:59Z","lastTransitionTime":"2025-12-06T03:34:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:34:59 crc kubenswrapper[4980]: I1206 03:34:59.999679 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:34:59 crc kubenswrapper[4980]: I1206 03:34:59.999746 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:34:59 crc kubenswrapper[4980]: I1206 03:34:59.999768 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:34:59 crc kubenswrapper[4980]: I1206 03:34:59.999796 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:34:59 crc kubenswrapper[4980]: I1206 03:34:59.999818 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:34:59Z","lastTransitionTime":"2025-12-06T03:34:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:00 crc kubenswrapper[4980]: I1206 03:35:00.103463 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:00 crc kubenswrapper[4980]: I1206 03:35:00.103555 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:00 crc kubenswrapper[4980]: I1206 03:35:00.103575 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:00 crc kubenswrapper[4980]: I1206 03:35:00.103598 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:00 crc kubenswrapper[4980]: I1206 03:35:00.103615 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:00Z","lastTransitionTime":"2025-12-06T03:35:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:00 crc kubenswrapper[4980]: I1206 03:35:00.184345 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 03:35:00 crc kubenswrapper[4980]: E1206 03:35:00.184573 4980 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 03:35:00 crc kubenswrapper[4980]: I1206 03:35:00.184963 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 03:35:00 crc kubenswrapper[4980]: I1206 03:35:00.185010 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-mcprh" Dec 06 03:35:00 crc kubenswrapper[4980]: E1206 03:35:00.185078 4980 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 03:35:00 crc kubenswrapper[4980]: E1206 03:35:00.185288 4980 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-mcprh" podUID="e6e88c5e-a255-449a-ae38-ac4d73a8e984" Dec 06 03:35:00 crc kubenswrapper[4980]: I1206 03:35:00.206133 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:00 crc kubenswrapper[4980]: I1206 03:35:00.206339 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:00 crc kubenswrapper[4980]: I1206 03:35:00.206471 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:00 crc kubenswrapper[4980]: I1206 03:35:00.206498 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:00 crc kubenswrapper[4980]: I1206 03:35:00.206546 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:00Z","lastTransitionTime":"2025-12-06T03:35:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:00 crc kubenswrapper[4980]: I1206 03:35:00.310199 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:00 crc kubenswrapper[4980]: I1206 03:35:00.310251 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:00 crc kubenswrapper[4980]: I1206 03:35:00.310268 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:00 crc kubenswrapper[4980]: I1206 03:35:00.310293 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:00 crc kubenswrapper[4980]: I1206 03:35:00.310309 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:00Z","lastTransitionTime":"2025-12-06T03:35:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:00 crc kubenswrapper[4980]: I1206 03:35:00.413908 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:00 crc kubenswrapper[4980]: I1206 03:35:00.413950 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:00 crc kubenswrapper[4980]: I1206 03:35:00.413959 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:00 crc kubenswrapper[4980]: I1206 03:35:00.413974 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:00 crc kubenswrapper[4980]: I1206 03:35:00.413984 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:00Z","lastTransitionTime":"2025-12-06T03:35:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:00 crc kubenswrapper[4980]: I1206 03:35:00.517664 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:00 crc kubenswrapper[4980]: I1206 03:35:00.517738 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:00 crc kubenswrapper[4980]: I1206 03:35:00.517755 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:00 crc kubenswrapper[4980]: I1206 03:35:00.517781 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:00 crc kubenswrapper[4980]: I1206 03:35:00.517798 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:00Z","lastTransitionTime":"2025-12-06T03:35:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:00 crc kubenswrapper[4980]: I1206 03:35:00.621887 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:00 crc kubenswrapper[4980]: I1206 03:35:00.621985 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:00 crc kubenswrapper[4980]: I1206 03:35:00.622011 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:00 crc kubenswrapper[4980]: I1206 03:35:00.622043 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:00 crc kubenswrapper[4980]: I1206 03:35:00.622067 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:00Z","lastTransitionTime":"2025-12-06T03:35:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:00 crc kubenswrapper[4980]: I1206 03:35:00.725086 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:00 crc kubenswrapper[4980]: I1206 03:35:00.725151 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:00 crc kubenswrapper[4980]: I1206 03:35:00.725168 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:00 crc kubenswrapper[4980]: I1206 03:35:00.725195 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:00 crc kubenswrapper[4980]: I1206 03:35:00.725213 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:00Z","lastTransitionTime":"2025-12-06T03:35:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:00 crc kubenswrapper[4980]: I1206 03:35:00.834315 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:00 crc kubenswrapper[4980]: I1206 03:35:00.834375 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:00 crc kubenswrapper[4980]: I1206 03:35:00.834389 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:00 crc kubenswrapper[4980]: I1206 03:35:00.834411 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:00 crc kubenswrapper[4980]: I1206 03:35:00.834423 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:00Z","lastTransitionTime":"2025-12-06T03:35:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:00 crc kubenswrapper[4980]: I1206 03:35:00.937246 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:00 crc kubenswrapper[4980]: I1206 03:35:00.937319 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:00 crc kubenswrapper[4980]: I1206 03:35:00.937343 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:00 crc kubenswrapper[4980]: I1206 03:35:00.937372 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:00 crc kubenswrapper[4980]: I1206 03:35:00.937395 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:00Z","lastTransitionTime":"2025-12-06T03:35:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:01 crc kubenswrapper[4980]: I1206 03:35:01.040804 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:01 crc kubenswrapper[4980]: I1206 03:35:01.040859 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:01 crc kubenswrapper[4980]: I1206 03:35:01.040878 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:01 crc kubenswrapper[4980]: I1206 03:35:01.040910 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:01 crc kubenswrapper[4980]: I1206 03:35:01.040930 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:01Z","lastTransitionTime":"2025-12-06T03:35:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:01 crc kubenswrapper[4980]: I1206 03:35:01.143847 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:01 crc kubenswrapper[4980]: I1206 03:35:01.143886 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:01 crc kubenswrapper[4980]: I1206 03:35:01.143894 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:01 crc kubenswrapper[4980]: I1206 03:35:01.143909 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:01 crc kubenswrapper[4980]: I1206 03:35:01.143919 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:01Z","lastTransitionTime":"2025-12-06T03:35:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:01 crc kubenswrapper[4980]: I1206 03:35:01.183418 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 03:35:01 crc kubenswrapper[4980]: E1206 03:35:01.183590 4980 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 03:35:01 crc kubenswrapper[4980]: I1206 03:35:01.197165 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"64278250-0b19-4591-b06e-6c96652ee92a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5beac3fd93b448506c0a110d70ef5faf42efa18b16abe6b3d4eb7b9c4233ad93\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ae66052702fcfd94f0123f4fbbe9efed37eb04609269a2bc951f95669de16601\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://173903f3554d7ac9158493498a016d618d92f46da1500408834b3c24249cf117\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1f596bfd5321ca225f4be03d4cf229c31a52def9731c0e177ab3a0bc010380e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1f596bfd5321ca225f4be03d4cf229c31a52def9731c0e177ab3a0bc010380e1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:02Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:01Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:35:01Z is after 2025-08-24T17:21:41Z" Dec 06 03:35:01 crc kubenswrapper[4980]: I1206 03:35:01.215102 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:22Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:35:01Z is after 2025-08-24T17:21:41Z" Dec 06 03:35:01 crc kubenswrapper[4980]: I1206 03:35:01.227758 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://334ea6a273fd7d268906c9f4a6dbf6f4df4fcdbf38f3d7d248cb7d6cf8537794\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ec44dcd5e4bf0a0466bb586ebd61259364dbbfe22499f4cbff854e3bd153082\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:35:01Z is after 2025-08-24T17:21:41Z" Dec 06 03:35:01 crc kubenswrapper[4980]: I1206 03:35:01.240365 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:22Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:35:01Z is after 2025-08-24T17:21:41Z" Dec 06 03:35:01 crc kubenswrapper[4980]: I1206 03:35:01.246217 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:01 crc kubenswrapper[4980]: I1206 03:35:01.246257 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:01 crc kubenswrapper[4980]: I1206 03:35:01.246267 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:01 crc kubenswrapper[4980]: I1206 03:35:01.246302 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:01 crc kubenswrapper[4980]: I1206 03:35:01.246311 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:01Z","lastTransitionTime":"2025-12-06T03:35:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:01 crc kubenswrapper[4980]: I1206 03:35:01.249387 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-sv5sz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"873977aa-4eb0-4526-bc0b-757ad2dea2a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01f0cc855c822a71ea8ef6a7ce6e25747ff9956691735973f935c4a28672a7a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lg7tx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:23Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-sv5sz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:35:01Z is after 2025-08-24T17:21:41Z" Dec 06 03:35:01 crc kubenswrapper[4980]: I1206 03:35:01.269611 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5w4fr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5af0e768-3c9c-4401-ab56-57516cd3170e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:23Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:23Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://768b8e1fe1a00a8467d744da5ad3f61006b0a3612c96f8e4d0ee6fcb93195f9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfzrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6342903b318317ee8d84c2f5d467cb25c811a789be0d6ab0e0cba37b5611e16d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfzrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d620685574db7594132ae24a91ec0a93dce20a9a289540deebb39650b8a1816\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfzrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c67ec8f9c1764fcdf215d52822c2d4f0437dc7f056f05d708767121f6c1f3515\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfzrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f35beae5339eeaf26494ed678e3c1bd02349eae2a4b4615325b17923f3b653b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfzrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a940ec6ee523313728f7e8ae934b7e0b0cff82771e3e64723309db1fae013c3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfzrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fae428d1a6eaeb2e6ddd4a5b1a1170e230635739ea0b39743e565e47fb7ee6e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fae428d1a6eaeb2e6ddd4a5b1a1170e230635739ea0b39743e565e47fb7ee6e4\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-06T03:34:53Z\\\",\\\"message\\\":\\\"/pkg/client/informers/externalversions/factory.go:141\\\\nI1206 03:34:53.908007 6569 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1206 03:34:53.908049 6569 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1206 03:34:53.908094 6569 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1206 03:34:53.908109 6569 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1206 03:34:53.908160 6569 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1206 03:34:53.908208 6569 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1206 03:34:53.908256 6569 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1206 03:34:53.908286 6569 factory.go:656] Stopping watch factory\\\\nI1206 03:34:53.908308 6569 ovnkube.go:599] Stopped ovnkube\\\\nI1206 03:34:53.908349 6569 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1206 03:34:53.908364 6569 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1206 03:34:53.908372 6569 handler.go:208] Removed *v1.Node event handler 2\\\\nI1206 03:34:53.908385 6569 handler.go:208] Removed *v1.Node event handler 7\\\\nI1206 03:34:53.908397 6569 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1206 03:34:53.908415 6569 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1206 03:34:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:52Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-5w4fr_openshift-ovn-kubernetes(5af0e768-3c9c-4401-ab56-57516cd3170e)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfzrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe261f4322b85702181d7473ab2a9a4d2d4b869d586c3091f811b12a81a63aae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfzrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d876df04b5ab0ba07a5cbcafbe4648403c5014c59eb7d40adee2657c3a255473\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d876df04b5ab0ba07a5cbcafbe4648403c5014c59eb7d40adee2657c3a255473\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfzrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:23Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-5w4fr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:35:01Z is after 2025-08-24T17:21:41Z" Dec 06 03:35:01 crc kubenswrapper[4980]: I1206 03:35:01.292438 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cab608b3-a02f-475b-9aae-9b8f5d878de9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f7fcc44b42bcbb55509048e3a566bb557333c4a83e33d7b0fba8eb514317909\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c22e7dd8886c2dfbb742900ed86ec7b0c2edf9a1363e74f6666961d928f06c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://473e792b2d0a9644854ee0581c6120fc9c989bc342899ef72b1d056080fd4694\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ee8e936dfc39361d6f99460910591ca8c901a20f0d274bbb6aaebfa8a00ee06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://77bc12768b018625ba11a6bfd66688f926c4b7f3bc1d0e41c6551a18b485ecf3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://27cf07cd96893900da4afecd0d56fd9269bc60c1d23246b4685fcd8756357c19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://27cf07cd96893900da4afecd0d56fd9269bc60c1d23246b4685fcd8756357c19\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c49feb24566a8284378670fa0390710d38c237acc94ffb654f90f3fa4626e4c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c49feb24566a8284378670fa0390710d38c237acc94ffb654f90f3fa4626e4c8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:03Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://bb74d1c3d752a379fa394c450217badb45511fbcafad62f7b618131cab94b29e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bb74d1c3d752a379fa394c450217badb45511fbcafad62f7b618131cab94b29e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:01Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:35:01Z is after 2025-08-24T17:21:41Z" Dec 06 03:35:01 crc kubenswrapper[4980]: I1206 03:35:01.306929 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://502c727d6baa31f400bee14c34b19abbe38af873f28f6a82a43547431755d981\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:35:01Z is after 2025-08-24T17:21:41Z" Dec 06 03:35:01 crc kubenswrapper[4980]: I1206 03:35:01.325469 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-dx44q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5bac22a-512c-49d0-90db-47de6aaaefd0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://97469dea2cd5393da0acf784ccc1969c8a4d07da7d51b14cdc955013ee9710ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f7mdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3b5dab476c7f227a6e4a2eae8469432667fb9fad345d9419ba59feab9fd385e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c3b5dab476c7f227a6e4a2eae8469432667fb9fad345d9419ba59feab9fd385e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f7mdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e921499231cbdf26b8918c98f9ad8c322fd7848865caeef048470de3e2c0efba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e921499231cbdf26b8918c98f9ad8c322fd7848865caeef048470de3e2c0efba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f7mdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7cbfe7d8cdb89916c238296592fab45b9e92e2c585baa6d67230decd1c84c5bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7cbfe7d8cdb89916c238296592fab45b9e92e2c585baa6d67230decd1c84c5bb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f7mdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f45cbf23b6b0d208cefc4b493f901bab9d71ccf2f1456507d54fb1f30c4fd48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7f45cbf23b6b0d208cefc4b493f901bab9d71ccf2f1456507d54fb1f30c4fd48\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f7mdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://344ce4b145ff021fcaf163803d8ad8cc6e9565bf7f0992e2653c43db663d8411\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://344ce4b145ff021fcaf163803d8ad8cc6e9565bf7f0992e2653c43db663d8411\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f7mdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://619e56a9937dd6376ab471ba16da30e5d77f443f5bd33610bea5ec85cd0446ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://619e56a9937dd6376ab471ba16da30e5d77f443f5bd33610bea5ec85cd0446ef\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f7mdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:23Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-dx44q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:35:01Z is after 2025-08-24T17:21:41Z" Dec 06 03:35:01 crc kubenswrapper[4980]: I1206 03:35:01.337749 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-r5zfb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"320f44d1-a671-4a91-b328-a8b0fdd8f23a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://92393ecaaef8cfcfd01837671f8cc89803dd570b637b4a373b40b7f487788826\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zhm5t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52e8db644463794041978ced2c0591c13e0ada8e5f1a78a94a0a665218a37b5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zhm5t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:23Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-r5zfb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:35:01Z is after 2025-08-24T17:21:41Z" Dec 06 03:35:01 crc kubenswrapper[4980]: I1206 03:35:01.350397 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:01 crc kubenswrapper[4980]: I1206 03:35:01.350445 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:01 crc kubenswrapper[4980]: I1206 03:35:01.350479 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:01 crc kubenswrapper[4980]: I1206 03:35:01.350496 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:01 crc kubenswrapper[4980]: I1206 03:35:01.350529 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:01Z","lastTransitionTime":"2025-12-06T03:35:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:01 crc kubenswrapper[4980]: I1206 03:35:01.350946 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-cvzg8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f1cde706-268d-4d26-818f-2f42bfb37b5f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://86f9a77acae360740be827eb0d1e1516e8a49cb9b197b6f5f9468ac70a101158\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rk86l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:29Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-cvzg8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:35:01Z is after 2025-08-24T17:21:41Z" Dec 06 03:35:01 crc kubenswrapper[4980]: I1206 03:35:01.367108 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:35:01Z is after 2025-08-24T17:21:41Z" Dec 06 03:35:01 crc kubenswrapper[4980]: I1206 03:35:01.384055 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4ef01e7d9809fd9cfc46e3a58b07159cdc6f5a3c210d1c9efc49bca27b30858\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:35:01Z is after 2025-08-24T17:21:41Z" Dec 06 03:35:01 crc kubenswrapper[4980]: I1206 03:35:01.402871 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-kq4z8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f96c1f6c-55e8-4d86-bff3-5381581b7022\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://799498681afa7653408bc23b1e82d75c16e3c809d2ee7f322780e20c537eb710\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bdx7g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:23Z\\\"}}\" for pod \"openshift-multus\"/\"multus-kq4z8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:35:01Z is after 2025-08-24T17:21:41Z" Dec 06 03:35:01 crc kubenswrapper[4980]: I1206 03:35:01.418755 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-6kxfq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"951b6deb-bcea-4332-8ae8-c23ac7aaef36\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://423d30c3e65b04bdf6ff5ef53d8a7575a434c05e28489370a7370381ebc29de9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b8htf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://282df50dd9653c1cbe97fc2041d9e9ea91a992661682652aafc12e566d8f1b81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b8htf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:37Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-6kxfq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:35:01Z is after 2025-08-24T17:21:41Z" Dec 06 03:35:01 crc kubenswrapper[4980]: I1206 03:35:01.429947 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-mcprh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e6e88c5e-a255-449a-ae38-ac4d73a8e984\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g7xz9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g7xz9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:38Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-mcprh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:35:01Z is after 2025-08-24T17:21:41Z" Dec 06 03:35:01 crc kubenswrapper[4980]: I1206 03:35:01.447262 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7f958150-549d-4c80-8a7e-23c39e4f9500\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c28b6895b14e73f4f3eb7a8ce93076c05092ba9b58ea0e3e879c65bcfa227a21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://76cff8a2d652df38b04eefc053832baf4cbb5db778ff5e525b18c2825141b118\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://206f6f19fe3197f5d6798cc86c5473bf2decb02a2860ee7ffb5efb5f60952e8c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bafaa9670b05671fb57b658c2960bf3ee32c84e52f86a3a3bbab01d5a38ff0c4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://773a0bc2aab73ffc2717e987543defc9b786cc22300aa54b4ba4f3454504e331\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-06T03:34:22Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1206 03:34:15.305330 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1206 03:34:15.307249 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1592846546/tls.crt::/tmp/serving-cert-1592846546/tls.key\\\\\\\"\\\\nI1206 03:34:21.814037 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1206 03:34:21.816756 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1206 03:34:21.816780 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1206 03:34:21.816806 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1206 03:34:21.816813 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1206 03:34:21.831301 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1206 03:34:21.831311 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1206 03:34:21.831324 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 03:34:21.831330 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 03:34:21.831335 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1206 03:34:21.831339 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1206 03:34:21.831342 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1206 03:34:21.831346 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1206 03:34:21.911269 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:04Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d40e0e7446f8416bf9da94fc53c01be937f9a51a11eb101157d2ddd4597720d2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:04Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2cba89f2f1bed96d59f183628d900d9b72c91eb472e273c8bed0babcab485910\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2cba89f2f1bed96d59f183628d900d9b72c91eb472e273c8bed0babcab485910\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:01Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:35:01Z is after 2025-08-24T17:21:41Z" Dec 06 03:35:01 crc kubenswrapper[4980]: I1206 03:35:01.453188 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:01 crc kubenswrapper[4980]: I1206 03:35:01.453270 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:01 crc kubenswrapper[4980]: I1206 03:35:01.453297 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:01 crc kubenswrapper[4980]: I1206 03:35:01.453331 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:01 crc kubenswrapper[4980]: I1206 03:35:01.453356 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:01Z","lastTransitionTime":"2025-12-06T03:35:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:01 crc kubenswrapper[4980]: I1206 03:35:01.463675 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6ed68f66-1cf6-465a-8f66-76d5af6a44d8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://725e84f6bdfb722a8ce69158d03ea4c66a9f087642d55cb9651e003c81be0c8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://71f7b492ea07cdf7dc65fc48184ec1e64914eb1811bbfdddc0fb14c8ad6cb296\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b543aa4ee03d5a14909b67d876c5675c02e1df2a89a3c9ac8d3fb587fea53d8f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://32af3427e872a70cb3c2357116f04cabc8d3d9b69e7306b2e0de2ae22f9469d4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:01Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:35:01Z is after 2025-08-24T17:21:41Z" Dec 06 03:35:01 crc kubenswrapper[4980]: I1206 03:35:01.556917 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:01 crc kubenswrapper[4980]: I1206 03:35:01.556966 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:01 crc kubenswrapper[4980]: I1206 03:35:01.556979 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:01 crc kubenswrapper[4980]: I1206 03:35:01.556997 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:01 crc kubenswrapper[4980]: I1206 03:35:01.557010 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:01Z","lastTransitionTime":"2025-12-06T03:35:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:01 crc kubenswrapper[4980]: I1206 03:35:01.659483 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:01 crc kubenswrapper[4980]: I1206 03:35:01.659614 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:01 crc kubenswrapper[4980]: I1206 03:35:01.659640 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:01 crc kubenswrapper[4980]: I1206 03:35:01.659666 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:01 crc kubenswrapper[4980]: I1206 03:35:01.659685 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:01Z","lastTransitionTime":"2025-12-06T03:35:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:01 crc kubenswrapper[4980]: I1206 03:35:01.763498 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:01 crc kubenswrapper[4980]: I1206 03:35:01.763606 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:01 crc kubenswrapper[4980]: I1206 03:35:01.763629 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:01 crc kubenswrapper[4980]: I1206 03:35:01.763661 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:01 crc kubenswrapper[4980]: I1206 03:35:01.763683 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:01Z","lastTransitionTime":"2025-12-06T03:35:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:01 crc kubenswrapper[4980]: I1206 03:35:01.867460 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:01 crc kubenswrapper[4980]: I1206 03:35:01.867563 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:01 crc kubenswrapper[4980]: I1206 03:35:01.867589 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:01 crc kubenswrapper[4980]: I1206 03:35:01.867619 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:01 crc kubenswrapper[4980]: I1206 03:35:01.867642 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:01Z","lastTransitionTime":"2025-12-06T03:35:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:01 crc kubenswrapper[4980]: I1206 03:35:01.970755 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:01 crc kubenswrapper[4980]: I1206 03:35:01.970805 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:01 crc kubenswrapper[4980]: I1206 03:35:01.970822 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:01 crc kubenswrapper[4980]: I1206 03:35:01.970844 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:01 crc kubenswrapper[4980]: I1206 03:35:01.970862 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:01Z","lastTransitionTime":"2025-12-06T03:35:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:02 crc kubenswrapper[4980]: I1206 03:35:02.073647 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:02 crc kubenswrapper[4980]: I1206 03:35:02.073691 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:02 crc kubenswrapper[4980]: I1206 03:35:02.073703 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:02 crc kubenswrapper[4980]: I1206 03:35:02.073720 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:02 crc kubenswrapper[4980]: I1206 03:35:02.073732 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:02Z","lastTransitionTime":"2025-12-06T03:35:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:02 crc kubenswrapper[4980]: I1206 03:35:02.177394 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:02 crc kubenswrapper[4980]: I1206 03:35:02.177456 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:02 crc kubenswrapper[4980]: I1206 03:35:02.177472 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:02 crc kubenswrapper[4980]: I1206 03:35:02.177492 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:02 crc kubenswrapper[4980]: I1206 03:35:02.177552 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:02Z","lastTransitionTime":"2025-12-06T03:35:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:02 crc kubenswrapper[4980]: I1206 03:35:02.183753 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 03:35:02 crc kubenswrapper[4980]: I1206 03:35:02.183782 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-mcprh" Dec 06 03:35:02 crc kubenswrapper[4980]: E1206 03:35:02.183877 4980 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 03:35:02 crc kubenswrapper[4980]: I1206 03:35:02.183806 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 03:35:02 crc kubenswrapper[4980]: E1206 03:35:02.183973 4980 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-mcprh" podUID="e6e88c5e-a255-449a-ae38-ac4d73a8e984" Dec 06 03:35:02 crc kubenswrapper[4980]: E1206 03:35:02.184113 4980 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 03:35:02 crc kubenswrapper[4980]: I1206 03:35:02.281082 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:02 crc kubenswrapper[4980]: I1206 03:35:02.281161 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:02 crc kubenswrapper[4980]: I1206 03:35:02.281184 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:02 crc kubenswrapper[4980]: I1206 03:35:02.281212 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:02 crc kubenswrapper[4980]: I1206 03:35:02.281232 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:02Z","lastTransitionTime":"2025-12-06T03:35:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:02 crc kubenswrapper[4980]: I1206 03:35:02.385065 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:02 crc kubenswrapper[4980]: I1206 03:35:02.385164 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:02 crc kubenswrapper[4980]: I1206 03:35:02.385183 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:02 crc kubenswrapper[4980]: I1206 03:35:02.385243 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:02 crc kubenswrapper[4980]: I1206 03:35:02.385260 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:02Z","lastTransitionTime":"2025-12-06T03:35:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:02 crc kubenswrapper[4980]: I1206 03:35:02.488915 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:02 crc kubenswrapper[4980]: I1206 03:35:02.489001 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:02 crc kubenswrapper[4980]: I1206 03:35:02.489023 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:02 crc kubenswrapper[4980]: I1206 03:35:02.489051 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:02 crc kubenswrapper[4980]: I1206 03:35:02.489069 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:02Z","lastTransitionTime":"2025-12-06T03:35:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:02 crc kubenswrapper[4980]: I1206 03:35:02.592120 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:02 crc kubenswrapper[4980]: I1206 03:35:02.592173 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:02 crc kubenswrapper[4980]: I1206 03:35:02.592197 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:02 crc kubenswrapper[4980]: I1206 03:35:02.592219 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:02 crc kubenswrapper[4980]: I1206 03:35:02.592235 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:02Z","lastTransitionTime":"2025-12-06T03:35:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:02 crc kubenswrapper[4980]: I1206 03:35:02.695348 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:02 crc kubenswrapper[4980]: I1206 03:35:02.695408 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:02 crc kubenswrapper[4980]: I1206 03:35:02.695420 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:02 crc kubenswrapper[4980]: I1206 03:35:02.695437 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:02 crc kubenswrapper[4980]: I1206 03:35:02.695452 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:02Z","lastTransitionTime":"2025-12-06T03:35:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:02 crc kubenswrapper[4980]: I1206 03:35:02.799287 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:02 crc kubenswrapper[4980]: I1206 03:35:02.799329 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:02 crc kubenswrapper[4980]: I1206 03:35:02.799338 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:02 crc kubenswrapper[4980]: I1206 03:35:02.799351 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:02 crc kubenswrapper[4980]: I1206 03:35:02.799360 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:02Z","lastTransitionTime":"2025-12-06T03:35:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:02 crc kubenswrapper[4980]: I1206 03:35:02.903335 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:02 crc kubenswrapper[4980]: I1206 03:35:02.903430 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:02 crc kubenswrapper[4980]: I1206 03:35:02.903457 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:02 crc kubenswrapper[4980]: I1206 03:35:02.903558 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:02 crc kubenswrapper[4980]: I1206 03:35:02.903603 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:02Z","lastTransitionTime":"2025-12-06T03:35:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:03 crc kubenswrapper[4980]: I1206 03:35:03.006698 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:03 crc kubenswrapper[4980]: I1206 03:35:03.006778 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:03 crc kubenswrapper[4980]: I1206 03:35:03.006794 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:03 crc kubenswrapper[4980]: I1206 03:35:03.006842 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:03 crc kubenswrapper[4980]: I1206 03:35:03.006858 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:03Z","lastTransitionTime":"2025-12-06T03:35:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:03 crc kubenswrapper[4980]: I1206 03:35:03.109157 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:03 crc kubenswrapper[4980]: I1206 03:35:03.109254 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:03 crc kubenswrapper[4980]: I1206 03:35:03.109317 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:03 crc kubenswrapper[4980]: I1206 03:35:03.109346 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:03 crc kubenswrapper[4980]: I1206 03:35:03.109411 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:03Z","lastTransitionTime":"2025-12-06T03:35:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:03 crc kubenswrapper[4980]: I1206 03:35:03.183599 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 03:35:03 crc kubenswrapper[4980]: E1206 03:35:03.183780 4980 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 03:35:03 crc kubenswrapper[4980]: I1206 03:35:03.212583 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:03 crc kubenswrapper[4980]: I1206 03:35:03.212663 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:03 crc kubenswrapper[4980]: I1206 03:35:03.212777 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:03 crc kubenswrapper[4980]: I1206 03:35:03.212813 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:03 crc kubenswrapper[4980]: I1206 03:35:03.212839 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:03Z","lastTransitionTime":"2025-12-06T03:35:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:03 crc kubenswrapper[4980]: I1206 03:35:03.316382 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:03 crc kubenswrapper[4980]: I1206 03:35:03.316452 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:03 crc kubenswrapper[4980]: I1206 03:35:03.316480 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:03 crc kubenswrapper[4980]: I1206 03:35:03.316548 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:03 crc kubenswrapper[4980]: I1206 03:35:03.316575 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:03Z","lastTransitionTime":"2025-12-06T03:35:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:03 crc kubenswrapper[4980]: I1206 03:35:03.418977 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:03 crc kubenswrapper[4980]: I1206 03:35:03.419042 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:03 crc kubenswrapper[4980]: I1206 03:35:03.419068 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:03 crc kubenswrapper[4980]: I1206 03:35:03.419098 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:03 crc kubenswrapper[4980]: I1206 03:35:03.419122 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:03Z","lastTransitionTime":"2025-12-06T03:35:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:03 crc kubenswrapper[4980]: I1206 03:35:03.522200 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:03 crc kubenswrapper[4980]: I1206 03:35:03.522748 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:03 crc kubenswrapper[4980]: I1206 03:35:03.522949 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:03 crc kubenswrapper[4980]: I1206 03:35:03.523192 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:03 crc kubenswrapper[4980]: I1206 03:35:03.523342 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:03Z","lastTransitionTime":"2025-12-06T03:35:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:03 crc kubenswrapper[4980]: I1206 03:35:03.627365 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:03 crc kubenswrapper[4980]: I1206 03:35:03.627413 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:03 crc kubenswrapper[4980]: I1206 03:35:03.627424 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:03 crc kubenswrapper[4980]: I1206 03:35:03.627439 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:03 crc kubenswrapper[4980]: I1206 03:35:03.627450 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:03Z","lastTransitionTime":"2025-12-06T03:35:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:03 crc kubenswrapper[4980]: I1206 03:35:03.730864 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:03 crc kubenswrapper[4980]: I1206 03:35:03.731306 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:03 crc kubenswrapper[4980]: I1206 03:35:03.731377 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:03 crc kubenswrapper[4980]: I1206 03:35:03.731493 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:03 crc kubenswrapper[4980]: I1206 03:35:03.731585 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:03Z","lastTransitionTime":"2025-12-06T03:35:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:03 crc kubenswrapper[4980]: I1206 03:35:03.835387 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:03 crc kubenswrapper[4980]: I1206 03:35:03.835460 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:03 crc kubenswrapper[4980]: I1206 03:35:03.835478 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:03 crc kubenswrapper[4980]: I1206 03:35:03.835499 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:03 crc kubenswrapper[4980]: I1206 03:35:03.835549 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:03Z","lastTransitionTime":"2025-12-06T03:35:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:03 crc kubenswrapper[4980]: I1206 03:35:03.939685 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:03 crc kubenswrapper[4980]: I1206 03:35:03.939748 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:03 crc kubenswrapper[4980]: I1206 03:35:03.939760 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:03 crc kubenswrapper[4980]: I1206 03:35:03.939810 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:03 crc kubenswrapper[4980]: I1206 03:35:03.939825 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:03Z","lastTransitionTime":"2025-12-06T03:35:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:04 crc kubenswrapper[4980]: I1206 03:35:04.044414 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:04 crc kubenswrapper[4980]: I1206 03:35:04.044871 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:04 crc kubenswrapper[4980]: I1206 03:35:04.044884 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:04 crc kubenswrapper[4980]: I1206 03:35:04.044902 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:04 crc kubenswrapper[4980]: I1206 03:35:04.044916 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:04Z","lastTransitionTime":"2025-12-06T03:35:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:04 crc kubenswrapper[4980]: I1206 03:35:04.148363 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:04 crc kubenswrapper[4980]: I1206 03:35:04.148446 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:04 crc kubenswrapper[4980]: I1206 03:35:04.148470 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:04 crc kubenswrapper[4980]: I1206 03:35:04.148505 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:04 crc kubenswrapper[4980]: I1206 03:35:04.148562 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:04Z","lastTransitionTime":"2025-12-06T03:35:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:04 crc kubenswrapper[4980]: I1206 03:35:04.183622 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 03:35:04 crc kubenswrapper[4980]: I1206 03:35:04.183694 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-mcprh" Dec 06 03:35:04 crc kubenswrapper[4980]: E1206 03:35:04.183766 4980 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 03:35:04 crc kubenswrapper[4980]: I1206 03:35:04.183696 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 03:35:04 crc kubenswrapper[4980]: E1206 03:35:04.183873 4980 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-mcprh" podUID="e6e88c5e-a255-449a-ae38-ac4d73a8e984" Dec 06 03:35:04 crc kubenswrapper[4980]: E1206 03:35:04.184026 4980 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 03:35:04 crc kubenswrapper[4980]: I1206 03:35:04.251862 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:04 crc kubenswrapper[4980]: I1206 03:35:04.251922 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:04 crc kubenswrapper[4980]: I1206 03:35:04.251934 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:04 crc kubenswrapper[4980]: I1206 03:35:04.251953 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:04 crc kubenswrapper[4980]: I1206 03:35:04.251970 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:04Z","lastTransitionTime":"2025-12-06T03:35:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:04 crc kubenswrapper[4980]: I1206 03:35:04.355836 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:04 crc kubenswrapper[4980]: I1206 03:35:04.355905 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:04 crc kubenswrapper[4980]: I1206 03:35:04.355923 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:04 crc kubenswrapper[4980]: I1206 03:35:04.355953 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:04 crc kubenswrapper[4980]: I1206 03:35:04.355973 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:04Z","lastTransitionTime":"2025-12-06T03:35:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:04 crc kubenswrapper[4980]: I1206 03:35:04.458941 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:04 crc kubenswrapper[4980]: I1206 03:35:04.459023 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:04 crc kubenswrapper[4980]: I1206 03:35:04.459056 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:04 crc kubenswrapper[4980]: I1206 03:35:04.459084 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:04 crc kubenswrapper[4980]: I1206 03:35:04.459101 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:04Z","lastTransitionTime":"2025-12-06T03:35:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:04 crc kubenswrapper[4980]: I1206 03:35:04.561785 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:04 crc kubenswrapper[4980]: I1206 03:35:04.561836 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:04 crc kubenswrapper[4980]: I1206 03:35:04.561859 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:04 crc kubenswrapper[4980]: I1206 03:35:04.561882 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:04 crc kubenswrapper[4980]: I1206 03:35:04.561894 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:04Z","lastTransitionTime":"2025-12-06T03:35:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:04 crc kubenswrapper[4980]: I1206 03:35:04.665125 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:04 crc kubenswrapper[4980]: I1206 03:35:04.665172 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:04 crc kubenswrapper[4980]: I1206 03:35:04.665183 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:04 crc kubenswrapper[4980]: I1206 03:35:04.665204 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:04 crc kubenswrapper[4980]: I1206 03:35:04.665217 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:04Z","lastTransitionTime":"2025-12-06T03:35:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:04 crc kubenswrapper[4980]: I1206 03:35:04.767402 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:04 crc kubenswrapper[4980]: I1206 03:35:04.767444 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:04 crc kubenswrapper[4980]: I1206 03:35:04.767455 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:04 crc kubenswrapper[4980]: I1206 03:35:04.767472 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:04 crc kubenswrapper[4980]: I1206 03:35:04.767486 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:04Z","lastTransitionTime":"2025-12-06T03:35:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:04 crc kubenswrapper[4980]: I1206 03:35:04.871220 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:04 crc kubenswrapper[4980]: I1206 03:35:04.871291 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:04 crc kubenswrapper[4980]: I1206 03:35:04.871308 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:04 crc kubenswrapper[4980]: I1206 03:35:04.871338 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:04 crc kubenswrapper[4980]: I1206 03:35:04.871356 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:04Z","lastTransitionTime":"2025-12-06T03:35:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:04 crc kubenswrapper[4980]: I1206 03:35:04.973917 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:04 crc kubenswrapper[4980]: I1206 03:35:04.973964 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:04 crc kubenswrapper[4980]: I1206 03:35:04.973974 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:04 crc kubenswrapper[4980]: I1206 03:35:04.973992 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:04 crc kubenswrapper[4980]: I1206 03:35:04.974003 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:04Z","lastTransitionTime":"2025-12-06T03:35:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:05 crc kubenswrapper[4980]: I1206 03:35:05.076768 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:05 crc kubenswrapper[4980]: I1206 03:35:05.076857 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:05 crc kubenswrapper[4980]: I1206 03:35:05.076873 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:05 crc kubenswrapper[4980]: I1206 03:35:05.076894 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:05 crc kubenswrapper[4980]: I1206 03:35:05.076907 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:05Z","lastTransitionTime":"2025-12-06T03:35:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:05 crc kubenswrapper[4980]: I1206 03:35:05.180531 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:05 crc kubenswrapper[4980]: I1206 03:35:05.180580 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:05 crc kubenswrapper[4980]: I1206 03:35:05.180593 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:05 crc kubenswrapper[4980]: I1206 03:35:05.180611 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:05 crc kubenswrapper[4980]: I1206 03:35:05.180622 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:05Z","lastTransitionTime":"2025-12-06T03:35:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:05 crc kubenswrapper[4980]: I1206 03:35:05.184143 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 03:35:05 crc kubenswrapper[4980]: E1206 03:35:05.184276 4980 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 03:35:05 crc kubenswrapper[4980]: I1206 03:35:05.283274 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:05 crc kubenswrapper[4980]: I1206 03:35:05.283323 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:05 crc kubenswrapper[4980]: I1206 03:35:05.283335 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:05 crc kubenswrapper[4980]: I1206 03:35:05.283358 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:05 crc kubenswrapper[4980]: I1206 03:35:05.283369 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:05Z","lastTransitionTime":"2025-12-06T03:35:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:05 crc kubenswrapper[4980]: I1206 03:35:05.385921 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:05 crc kubenswrapper[4980]: I1206 03:35:05.385954 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:05 crc kubenswrapper[4980]: I1206 03:35:05.385963 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:05 crc kubenswrapper[4980]: I1206 03:35:05.385975 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:05 crc kubenswrapper[4980]: I1206 03:35:05.385985 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:05Z","lastTransitionTime":"2025-12-06T03:35:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:05 crc kubenswrapper[4980]: I1206 03:35:05.488175 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:05 crc kubenswrapper[4980]: I1206 03:35:05.488225 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:05 crc kubenswrapper[4980]: I1206 03:35:05.488238 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:05 crc kubenswrapper[4980]: I1206 03:35:05.488256 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:05 crc kubenswrapper[4980]: I1206 03:35:05.488266 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:05Z","lastTransitionTime":"2025-12-06T03:35:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:05 crc kubenswrapper[4980]: I1206 03:35:05.590205 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:05 crc kubenswrapper[4980]: I1206 03:35:05.590264 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:05 crc kubenswrapper[4980]: I1206 03:35:05.590277 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:05 crc kubenswrapper[4980]: I1206 03:35:05.590302 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:05 crc kubenswrapper[4980]: I1206 03:35:05.590316 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:05Z","lastTransitionTime":"2025-12-06T03:35:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:05 crc kubenswrapper[4980]: I1206 03:35:05.693424 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:05 crc kubenswrapper[4980]: I1206 03:35:05.693473 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:05 crc kubenswrapper[4980]: I1206 03:35:05.693486 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:05 crc kubenswrapper[4980]: I1206 03:35:05.693504 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:05 crc kubenswrapper[4980]: I1206 03:35:05.693536 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:05Z","lastTransitionTime":"2025-12-06T03:35:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:05 crc kubenswrapper[4980]: I1206 03:35:05.796374 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:05 crc kubenswrapper[4980]: I1206 03:35:05.796419 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:05 crc kubenswrapper[4980]: I1206 03:35:05.796430 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:05 crc kubenswrapper[4980]: I1206 03:35:05.796446 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:05 crc kubenswrapper[4980]: I1206 03:35:05.796457 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:05Z","lastTransitionTime":"2025-12-06T03:35:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:05 crc kubenswrapper[4980]: I1206 03:35:05.898652 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:05 crc kubenswrapper[4980]: I1206 03:35:05.898698 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:05 crc kubenswrapper[4980]: I1206 03:35:05.898708 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:05 crc kubenswrapper[4980]: I1206 03:35:05.898724 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:05 crc kubenswrapper[4980]: I1206 03:35:05.898737 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:05Z","lastTransitionTime":"2025-12-06T03:35:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:06 crc kubenswrapper[4980]: I1206 03:35:06.001180 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:06 crc kubenswrapper[4980]: I1206 03:35:06.001225 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:06 crc kubenswrapper[4980]: I1206 03:35:06.001235 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:06 crc kubenswrapper[4980]: I1206 03:35:06.001250 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:06 crc kubenswrapper[4980]: I1206 03:35:06.001260 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:06Z","lastTransitionTime":"2025-12-06T03:35:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:06 crc kubenswrapper[4980]: I1206 03:35:06.103427 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:06 crc kubenswrapper[4980]: I1206 03:35:06.103460 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:06 crc kubenswrapper[4980]: I1206 03:35:06.103468 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:06 crc kubenswrapper[4980]: I1206 03:35:06.103482 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:06 crc kubenswrapper[4980]: I1206 03:35:06.103491 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:06Z","lastTransitionTime":"2025-12-06T03:35:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:06 crc kubenswrapper[4980]: I1206 03:35:06.183354 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 03:35:06 crc kubenswrapper[4980]: I1206 03:35:06.183354 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-mcprh" Dec 06 03:35:06 crc kubenswrapper[4980]: I1206 03:35:06.183363 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 03:35:06 crc kubenswrapper[4980]: E1206 03:35:06.183662 4980 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-mcprh" podUID="e6e88c5e-a255-449a-ae38-ac4d73a8e984" Dec 06 03:35:06 crc kubenswrapper[4980]: E1206 03:35:06.183746 4980 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 03:35:06 crc kubenswrapper[4980]: E1206 03:35:06.183800 4980 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 03:35:06 crc kubenswrapper[4980]: I1206 03:35:06.206166 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:06 crc kubenswrapper[4980]: I1206 03:35:06.206203 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:06 crc kubenswrapper[4980]: I1206 03:35:06.206212 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:06 crc kubenswrapper[4980]: I1206 03:35:06.206229 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:06 crc kubenswrapper[4980]: I1206 03:35:06.206242 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:06Z","lastTransitionTime":"2025-12-06T03:35:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:06 crc kubenswrapper[4980]: I1206 03:35:06.309059 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:06 crc kubenswrapper[4980]: I1206 03:35:06.309111 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:06 crc kubenswrapper[4980]: I1206 03:35:06.309125 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:06 crc kubenswrapper[4980]: I1206 03:35:06.309143 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:06 crc kubenswrapper[4980]: I1206 03:35:06.309156 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:06Z","lastTransitionTime":"2025-12-06T03:35:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:06 crc kubenswrapper[4980]: I1206 03:35:06.412627 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:06 crc kubenswrapper[4980]: I1206 03:35:06.412669 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:06 crc kubenswrapper[4980]: I1206 03:35:06.412680 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:06 crc kubenswrapper[4980]: I1206 03:35:06.412695 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:06 crc kubenswrapper[4980]: I1206 03:35:06.412706 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:06Z","lastTransitionTime":"2025-12-06T03:35:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:06 crc kubenswrapper[4980]: I1206 03:35:06.517142 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:06 crc kubenswrapper[4980]: I1206 03:35:06.517200 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:06 crc kubenswrapper[4980]: I1206 03:35:06.517230 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:06 crc kubenswrapper[4980]: I1206 03:35:06.517248 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:06 crc kubenswrapper[4980]: I1206 03:35:06.517261 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:06Z","lastTransitionTime":"2025-12-06T03:35:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:06 crc kubenswrapper[4980]: I1206 03:35:06.620048 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:06 crc kubenswrapper[4980]: I1206 03:35:06.620102 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:06 crc kubenswrapper[4980]: I1206 03:35:06.620115 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:06 crc kubenswrapper[4980]: I1206 03:35:06.620133 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:06 crc kubenswrapper[4980]: I1206 03:35:06.620145 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:06Z","lastTransitionTime":"2025-12-06T03:35:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:06 crc kubenswrapper[4980]: I1206 03:35:06.730124 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:06 crc kubenswrapper[4980]: I1206 03:35:06.730188 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:06 crc kubenswrapper[4980]: I1206 03:35:06.730201 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:06 crc kubenswrapper[4980]: I1206 03:35:06.730221 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:06 crc kubenswrapper[4980]: I1206 03:35:06.730235 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:06Z","lastTransitionTime":"2025-12-06T03:35:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:06 crc kubenswrapper[4980]: I1206 03:35:06.833657 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:06 crc kubenswrapper[4980]: I1206 03:35:06.833722 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:06 crc kubenswrapper[4980]: I1206 03:35:06.833742 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:06 crc kubenswrapper[4980]: I1206 03:35:06.833765 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:06 crc kubenswrapper[4980]: I1206 03:35:06.833781 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:06Z","lastTransitionTime":"2025-12-06T03:35:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:06 crc kubenswrapper[4980]: I1206 03:35:06.936058 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:06 crc kubenswrapper[4980]: I1206 03:35:06.936094 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:06 crc kubenswrapper[4980]: I1206 03:35:06.936103 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:06 crc kubenswrapper[4980]: I1206 03:35:06.936116 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:06 crc kubenswrapper[4980]: I1206 03:35:06.936124 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:06Z","lastTransitionTime":"2025-12-06T03:35:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:07 crc kubenswrapper[4980]: I1206 03:35:07.038901 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:07 crc kubenswrapper[4980]: I1206 03:35:07.038947 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:07 crc kubenswrapper[4980]: I1206 03:35:07.038958 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:07 crc kubenswrapper[4980]: I1206 03:35:07.038978 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:07 crc kubenswrapper[4980]: I1206 03:35:07.038990 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:07Z","lastTransitionTime":"2025-12-06T03:35:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:07 crc kubenswrapper[4980]: I1206 03:35:07.142122 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:07 crc kubenswrapper[4980]: I1206 03:35:07.142173 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:07 crc kubenswrapper[4980]: I1206 03:35:07.142181 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:07 crc kubenswrapper[4980]: I1206 03:35:07.142195 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:07 crc kubenswrapper[4980]: I1206 03:35:07.142205 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:07Z","lastTransitionTime":"2025-12-06T03:35:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:07 crc kubenswrapper[4980]: I1206 03:35:07.183653 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 03:35:07 crc kubenswrapper[4980]: E1206 03:35:07.183893 4980 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 03:35:07 crc kubenswrapper[4980]: I1206 03:35:07.244659 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:07 crc kubenswrapper[4980]: I1206 03:35:07.244713 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:07 crc kubenswrapper[4980]: I1206 03:35:07.244728 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:07 crc kubenswrapper[4980]: I1206 03:35:07.244747 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:07 crc kubenswrapper[4980]: I1206 03:35:07.244761 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:07Z","lastTransitionTime":"2025-12-06T03:35:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:07 crc kubenswrapper[4980]: I1206 03:35:07.347440 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:07 crc kubenswrapper[4980]: I1206 03:35:07.347499 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:07 crc kubenswrapper[4980]: I1206 03:35:07.347537 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:07 crc kubenswrapper[4980]: I1206 03:35:07.347558 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:07 crc kubenswrapper[4980]: I1206 03:35:07.347572 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:07Z","lastTransitionTime":"2025-12-06T03:35:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:07 crc kubenswrapper[4980]: I1206 03:35:07.449865 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:07 crc kubenswrapper[4980]: I1206 03:35:07.449906 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:07 crc kubenswrapper[4980]: I1206 03:35:07.449917 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:07 crc kubenswrapper[4980]: I1206 03:35:07.449932 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:07 crc kubenswrapper[4980]: I1206 03:35:07.449943 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:07Z","lastTransitionTime":"2025-12-06T03:35:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:07 crc kubenswrapper[4980]: I1206 03:35:07.552415 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:07 crc kubenswrapper[4980]: I1206 03:35:07.552478 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:07 crc kubenswrapper[4980]: I1206 03:35:07.552490 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:07 crc kubenswrapper[4980]: I1206 03:35:07.552550 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:07 crc kubenswrapper[4980]: I1206 03:35:07.552565 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:07Z","lastTransitionTime":"2025-12-06T03:35:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:07 crc kubenswrapper[4980]: I1206 03:35:07.655688 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:07 crc kubenswrapper[4980]: I1206 03:35:07.655736 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:07 crc kubenswrapper[4980]: I1206 03:35:07.655745 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:07 crc kubenswrapper[4980]: I1206 03:35:07.655761 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:07 crc kubenswrapper[4980]: I1206 03:35:07.655771 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:07Z","lastTransitionTime":"2025-12-06T03:35:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:07 crc kubenswrapper[4980]: I1206 03:35:07.758787 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:07 crc kubenswrapper[4980]: I1206 03:35:07.758836 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:07 crc kubenswrapper[4980]: I1206 03:35:07.758846 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:07 crc kubenswrapper[4980]: I1206 03:35:07.758862 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:07 crc kubenswrapper[4980]: I1206 03:35:07.758871 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:07Z","lastTransitionTime":"2025-12-06T03:35:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:07 crc kubenswrapper[4980]: I1206 03:35:07.861791 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:07 crc kubenswrapper[4980]: I1206 03:35:07.861851 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:07 crc kubenswrapper[4980]: I1206 03:35:07.861862 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:07 crc kubenswrapper[4980]: I1206 03:35:07.861882 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:07 crc kubenswrapper[4980]: I1206 03:35:07.861896 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:07Z","lastTransitionTime":"2025-12-06T03:35:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:07 crc kubenswrapper[4980]: I1206 03:35:07.965067 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:07 crc kubenswrapper[4980]: I1206 03:35:07.965118 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:07 crc kubenswrapper[4980]: I1206 03:35:07.965127 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:07 crc kubenswrapper[4980]: I1206 03:35:07.965142 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:07 crc kubenswrapper[4980]: I1206 03:35:07.965152 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:07Z","lastTransitionTime":"2025-12-06T03:35:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:08 crc kubenswrapper[4980]: I1206 03:35:08.067557 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:08 crc kubenswrapper[4980]: I1206 03:35:08.067622 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:08 crc kubenswrapper[4980]: I1206 03:35:08.067633 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:08 crc kubenswrapper[4980]: I1206 03:35:08.067660 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:08 crc kubenswrapper[4980]: I1206 03:35:08.067680 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:08Z","lastTransitionTime":"2025-12-06T03:35:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:08 crc kubenswrapper[4980]: I1206 03:35:08.169545 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:08 crc kubenswrapper[4980]: I1206 03:35:08.169593 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:08 crc kubenswrapper[4980]: I1206 03:35:08.169609 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:08 crc kubenswrapper[4980]: I1206 03:35:08.169626 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:08 crc kubenswrapper[4980]: I1206 03:35:08.169642 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:08Z","lastTransitionTime":"2025-12-06T03:35:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:08 crc kubenswrapper[4980]: I1206 03:35:08.183784 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-mcprh" Dec 06 03:35:08 crc kubenswrapper[4980]: I1206 03:35:08.183817 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 03:35:08 crc kubenswrapper[4980]: I1206 03:35:08.183784 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 03:35:08 crc kubenswrapper[4980]: E1206 03:35:08.183901 4980 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-mcprh" podUID="e6e88c5e-a255-449a-ae38-ac4d73a8e984" Dec 06 03:35:08 crc kubenswrapper[4980]: E1206 03:35:08.183980 4980 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 03:35:08 crc kubenswrapper[4980]: E1206 03:35:08.184041 4980 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 03:35:08 crc kubenswrapper[4980]: I1206 03:35:08.272195 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:08 crc kubenswrapper[4980]: I1206 03:35:08.272266 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:08 crc kubenswrapper[4980]: I1206 03:35:08.272290 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:08 crc kubenswrapper[4980]: I1206 03:35:08.272319 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:08 crc kubenswrapper[4980]: I1206 03:35:08.272343 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:08Z","lastTransitionTime":"2025-12-06T03:35:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:08 crc kubenswrapper[4980]: I1206 03:35:08.375882 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:08 crc kubenswrapper[4980]: I1206 03:35:08.375930 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:08 crc kubenswrapper[4980]: I1206 03:35:08.375942 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:08 crc kubenswrapper[4980]: I1206 03:35:08.375959 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:08 crc kubenswrapper[4980]: I1206 03:35:08.375970 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:08Z","lastTransitionTime":"2025-12-06T03:35:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:08 crc kubenswrapper[4980]: I1206 03:35:08.478557 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:08 crc kubenswrapper[4980]: I1206 03:35:08.478611 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:08 crc kubenswrapper[4980]: I1206 03:35:08.478632 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:08 crc kubenswrapper[4980]: I1206 03:35:08.478649 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:08 crc kubenswrapper[4980]: I1206 03:35:08.478659 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:08Z","lastTransitionTime":"2025-12-06T03:35:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:08 crc kubenswrapper[4980]: I1206 03:35:08.581741 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:08 crc kubenswrapper[4980]: I1206 03:35:08.581809 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:08 crc kubenswrapper[4980]: I1206 03:35:08.581823 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:08 crc kubenswrapper[4980]: I1206 03:35:08.581840 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:08 crc kubenswrapper[4980]: I1206 03:35:08.581852 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:08Z","lastTransitionTime":"2025-12-06T03:35:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:08 crc kubenswrapper[4980]: I1206 03:35:08.684844 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:08 crc kubenswrapper[4980]: I1206 03:35:08.684914 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:08 crc kubenswrapper[4980]: I1206 03:35:08.684926 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:08 crc kubenswrapper[4980]: I1206 03:35:08.684942 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:08 crc kubenswrapper[4980]: I1206 03:35:08.684953 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:08Z","lastTransitionTime":"2025-12-06T03:35:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:08 crc kubenswrapper[4980]: I1206 03:35:08.788204 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:08 crc kubenswrapper[4980]: I1206 03:35:08.788258 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:08 crc kubenswrapper[4980]: I1206 03:35:08.788272 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:08 crc kubenswrapper[4980]: I1206 03:35:08.788300 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:08 crc kubenswrapper[4980]: I1206 03:35:08.788311 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:08Z","lastTransitionTime":"2025-12-06T03:35:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:08 crc kubenswrapper[4980]: I1206 03:35:08.891042 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:08 crc kubenswrapper[4980]: I1206 03:35:08.891107 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:08 crc kubenswrapper[4980]: I1206 03:35:08.891130 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:08 crc kubenswrapper[4980]: I1206 03:35:08.891162 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:08 crc kubenswrapper[4980]: I1206 03:35:08.891184 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:08Z","lastTransitionTime":"2025-12-06T03:35:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:08 crc kubenswrapper[4980]: I1206 03:35:08.994815 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:08 crc kubenswrapper[4980]: I1206 03:35:08.994886 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:08 crc kubenswrapper[4980]: I1206 03:35:08.994911 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:08 crc kubenswrapper[4980]: I1206 03:35:08.994941 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:08 crc kubenswrapper[4980]: I1206 03:35:08.994958 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:08Z","lastTransitionTime":"2025-12-06T03:35:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:09 crc kubenswrapper[4980]: I1206 03:35:09.098055 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:09 crc kubenswrapper[4980]: I1206 03:35:09.098138 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:09 crc kubenswrapper[4980]: I1206 03:35:09.098162 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:09 crc kubenswrapper[4980]: I1206 03:35:09.098193 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:09 crc kubenswrapper[4980]: I1206 03:35:09.098220 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:09Z","lastTransitionTime":"2025-12-06T03:35:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:09 crc kubenswrapper[4980]: I1206 03:35:09.184040 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 03:35:09 crc kubenswrapper[4980]: E1206 03:35:09.184796 4980 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 03:35:09 crc kubenswrapper[4980]: I1206 03:35:09.185139 4980 scope.go:117] "RemoveContainer" containerID="fae428d1a6eaeb2e6ddd4a5b1a1170e230635739ea0b39743e565e47fb7ee6e4" Dec 06 03:35:09 crc kubenswrapper[4980]: E1206 03:35:09.185558 4980 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-5w4fr_openshift-ovn-kubernetes(5af0e768-3c9c-4401-ab56-57516cd3170e)\"" pod="openshift-ovn-kubernetes/ovnkube-node-5w4fr" podUID="5af0e768-3c9c-4401-ab56-57516cd3170e" Dec 06 03:35:09 crc kubenswrapper[4980]: I1206 03:35:09.201823 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:09 crc kubenswrapper[4980]: I1206 03:35:09.201882 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:09 crc kubenswrapper[4980]: I1206 03:35:09.201895 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:09 crc kubenswrapper[4980]: I1206 03:35:09.201912 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:09 crc kubenswrapper[4980]: I1206 03:35:09.201925 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:09Z","lastTransitionTime":"2025-12-06T03:35:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:09 crc kubenswrapper[4980]: I1206 03:35:09.304777 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:09 crc kubenswrapper[4980]: I1206 03:35:09.304838 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:09 crc kubenswrapper[4980]: I1206 03:35:09.304854 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:09 crc kubenswrapper[4980]: I1206 03:35:09.304875 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:09 crc kubenswrapper[4980]: I1206 03:35:09.304889 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:09Z","lastTransitionTime":"2025-12-06T03:35:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:09 crc kubenswrapper[4980]: I1206 03:35:09.412344 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:09 crc kubenswrapper[4980]: I1206 03:35:09.412407 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:09 crc kubenswrapper[4980]: I1206 03:35:09.412423 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:09 crc kubenswrapper[4980]: I1206 03:35:09.412444 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:09 crc kubenswrapper[4980]: I1206 03:35:09.412458 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:09Z","lastTransitionTime":"2025-12-06T03:35:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:09 crc kubenswrapper[4980]: I1206 03:35:09.457328 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:09 crc kubenswrapper[4980]: I1206 03:35:09.457384 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:09 crc kubenswrapper[4980]: I1206 03:35:09.457394 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:09 crc kubenswrapper[4980]: I1206 03:35:09.457408 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:09 crc kubenswrapper[4980]: I1206 03:35:09.457417 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:09Z","lastTransitionTime":"2025-12-06T03:35:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:09 crc kubenswrapper[4980]: E1206 03:35:09.471983 4980 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T03:35:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T03:35:09Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T03:35:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T03:35:09Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T03:35:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T03:35:09Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T03:35:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T03:35:09Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"9b4ed4f8-ed8d-447b-9e1f-13defbdcf0f7\\\",\\\"systemUUID\\\":\\\"a5b493aa-69de-4688-8ff6-1e5f49b7f014\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:35:09Z is after 2025-08-24T17:21:41Z" Dec 06 03:35:09 crc kubenswrapper[4980]: I1206 03:35:09.476594 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:09 crc kubenswrapper[4980]: I1206 03:35:09.476663 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:09 crc kubenswrapper[4980]: I1206 03:35:09.476678 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:09 crc kubenswrapper[4980]: I1206 03:35:09.476706 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:09 crc kubenswrapper[4980]: I1206 03:35:09.476724 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:09Z","lastTransitionTime":"2025-12-06T03:35:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:09 crc kubenswrapper[4980]: E1206 03:35:09.488910 4980 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T03:35:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T03:35:09Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T03:35:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T03:35:09Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T03:35:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T03:35:09Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T03:35:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T03:35:09Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"9b4ed4f8-ed8d-447b-9e1f-13defbdcf0f7\\\",\\\"systemUUID\\\":\\\"a5b493aa-69de-4688-8ff6-1e5f49b7f014\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:35:09Z is after 2025-08-24T17:21:41Z" Dec 06 03:35:09 crc kubenswrapper[4980]: I1206 03:35:09.494100 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:09 crc kubenswrapper[4980]: I1206 03:35:09.494140 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:09 crc kubenswrapper[4980]: I1206 03:35:09.494155 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:09 crc kubenswrapper[4980]: I1206 03:35:09.494174 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:09 crc kubenswrapper[4980]: I1206 03:35:09.494187 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:09Z","lastTransitionTime":"2025-12-06T03:35:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:09 crc kubenswrapper[4980]: E1206 03:35:09.507504 4980 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T03:35:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T03:35:09Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T03:35:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T03:35:09Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T03:35:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T03:35:09Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T03:35:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T03:35:09Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"9b4ed4f8-ed8d-447b-9e1f-13defbdcf0f7\\\",\\\"systemUUID\\\":\\\"a5b493aa-69de-4688-8ff6-1e5f49b7f014\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:35:09Z is after 2025-08-24T17:21:41Z" Dec 06 03:35:09 crc kubenswrapper[4980]: I1206 03:35:09.511502 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:09 crc kubenswrapper[4980]: I1206 03:35:09.511550 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:09 crc kubenswrapper[4980]: I1206 03:35:09.511639 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:09 crc kubenswrapper[4980]: I1206 03:35:09.511656 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:09 crc kubenswrapper[4980]: I1206 03:35:09.511669 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:09Z","lastTransitionTime":"2025-12-06T03:35:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:09 crc kubenswrapper[4980]: E1206 03:35:09.524773 4980 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T03:35:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T03:35:09Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T03:35:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T03:35:09Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T03:35:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T03:35:09Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T03:35:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T03:35:09Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"9b4ed4f8-ed8d-447b-9e1f-13defbdcf0f7\\\",\\\"systemUUID\\\":\\\"a5b493aa-69de-4688-8ff6-1e5f49b7f014\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:35:09Z is after 2025-08-24T17:21:41Z" Dec 06 03:35:09 crc kubenswrapper[4980]: I1206 03:35:09.528105 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:09 crc kubenswrapper[4980]: I1206 03:35:09.528136 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:09 crc kubenswrapper[4980]: I1206 03:35:09.528147 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:09 crc kubenswrapper[4980]: I1206 03:35:09.528162 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:09 crc kubenswrapper[4980]: I1206 03:35:09.528173 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:09Z","lastTransitionTime":"2025-12-06T03:35:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:09 crc kubenswrapper[4980]: E1206 03:35:09.540205 4980 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T03:35:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T03:35:09Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T03:35:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T03:35:09Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T03:35:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T03:35:09Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T03:35:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T03:35:09Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"9b4ed4f8-ed8d-447b-9e1f-13defbdcf0f7\\\",\\\"systemUUID\\\":\\\"a5b493aa-69de-4688-8ff6-1e5f49b7f014\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:35:09Z is after 2025-08-24T17:21:41Z" Dec 06 03:35:09 crc kubenswrapper[4980]: E1206 03:35:09.540402 4980 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 06 03:35:09 crc kubenswrapper[4980]: I1206 03:35:09.542405 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:09 crc kubenswrapper[4980]: I1206 03:35:09.542445 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:09 crc kubenswrapper[4980]: I1206 03:35:09.542456 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:09 crc kubenswrapper[4980]: I1206 03:35:09.542471 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:09 crc kubenswrapper[4980]: I1206 03:35:09.542482 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:09Z","lastTransitionTime":"2025-12-06T03:35:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:09 crc kubenswrapper[4980]: I1206 03:35:09.645159 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:09 crc kubenswrapper[4980]: I1206 03:35:09.645205 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:09 crc kubenswrapper[4980]: I1206 03:35:09.645218 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:09 crc kubenswrapper[4980]: I1206 03:35:09.645234 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:09 crc kubenswrapper[4980]: I1206 03:35:09.645245 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:09Z","lastTransitionTime":"2025-12-06T03:35:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:09 crc kubenswrapper[4980]: I1206 03:35:09.748114 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:09 crc kubenswrapper[4980]: I1206 03:35:09.748151 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:09 crc kubenswrapper[4980]: I1206 03:35:09.748163 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:09 crc kubenswrapper[4980]: I1206 03:35:09.748179 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:09 crc kubenswrapper[4980]: I1206 03:35:09.748190 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:09Z","lastTransitionTime":"2025-12-06T03:35:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:09 crc kubenswrapper[4980]: I1206 03:35:09.850957 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:09 crc kubenswrapper[4980]: I1206 03:35:09.851009 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:09 crc kubenswrapper[4980]: I1206 03:35:09.851020 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:09 crc kubenswrapper[4980]: I1206 03:35:09.851036 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:09 crc kubenswrapper[4980]: I1206 03:35:09.851046 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:09Z","lastTransitionTime":"2025-12-06T03:35:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:09 crc kubenswrapper[4980]: I1206 03:35:09.953938 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:09 crc kubenswrapper[4980]: I1206 03:35:09.953995 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:09 crc kubenswrapper[4980]: I1206 03:35:09.954005 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:09 crc kubenswrapper[4980]: I1206 03:35:09.954018 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:09 crc kubenswrapper[4980]: I1206 03:35:09.954029 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:09Z","lastTransitionTime":"2025-12-06T03:35:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:10 crc kubenswrapper[4980]: I1206 03:35:10.057631 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:10 crc kubenswrapper[4980]: I1206 03:35:10.057689 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:10 crc kubenswrapper[4980]: I1206 03:35:10.057703 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:10 crc kubenswrapper[4980]: I1206 03:35:10.057723 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:10 crc kubenswrapper[4980]: I1206 03:35:10.057746 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:10Z","lastTransitionTime":"2025-12-06T03:35:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:10 crc kubenswrapper[4980]: I1206 03:35:10.161128 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:10 crc kubenswrapper[4980]: I1206 03:35:10.161202 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:10 crc kubenswrapper[4980]: I1206 03:35:10.161220 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:10 crc kubenswrapper[4980]: I1206 03:35:10.161246 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:10 crc kubenswrapper[4980]: I1206 03:35:10.161269 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:10Z","lastTransitionTime":"2025-12-06T03:35:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:10 crc kubenswrapper[4980]: I1206 03:35:10.183830 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 03:35:10 crc kubenswrapper[4980]: I1206 03:35:10.183932 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-mcprh" Dec 06 03:35:10 crc kubenswrapper[4980]: I1206 03:35:10.183974 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 03:35:10 crc kubenswrapper[4980]: E1206 03:35:10.184544 4980 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-mcprh" podUID="e6e88c5e-a255-449a-ae38-ac4d73a8e984" Dec 06 03:35:10 crc kubenswrapper[4980]: E1206 03:35:10.184322 4980 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 03:35:10 crc kubenswrapper[4980]: E1206 03:35:10.184705 4980 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 03:35:10 crc kubenswrapper[4980]: I1206 03:35:10.264393 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:10 crc kubenswrapper[4980]: I1206 03:35:10.264425 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:10 crc kubenswrapper[4980]: I1206 03:35:10.264433 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:10 crc kubenswrapper[4980]: I1206 03:35:10.264447 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:10 crc kubenswrapper[4980]: I1206 03:35:10.264455 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:10Z","lastTransitionTime":"2025-12-06T03:35:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:10 crc kubenswrapper[4980]: I1206 03:35:10.367324 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:10 crc kubenswrapper[4980]: I1206 03:35:10.367821 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:10 crc kubenswrapper[4980]: I1206 03:35:10.367996 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:10 crc kubenswrapper[4980]: I1206 03:35:10.368168 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:10 crc kubenswrapper[4980]: I1206 03:35:10.368314 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:10Z","lastTransitionTime":"2025-12-06T03:35:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:10 crc kubenswrapper[4980]: I1206 03:35:10.471339 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:10 crc kubenswrapper[4980]: I1206 03:35:10.471399 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:10 crc kubenswrapper[4980]: I1206 03:35:10.471412 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:10 crc kubenswrapper[4980]: I1206 03:35:10.471437 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:10 crc kubenswrapper[4980]: I1206 03:35:10.471449 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:10Z","lastTransitionTime":"2025-12-06T03:35:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:10 crc kubenswrapper[4980]: I1206 03:35:10.554111 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/e6e88c5e-a255-449a-ae38-ac4d73a8e984-metrics-certs\") pod \"network-metrics-daemon-mcprh\" (UID: \"e6e88c5e-a255-449a-ae38-ac4d73a8e984\") " pod="openshift-multus/network-metrics-daemon-mcprh" Dec 06 03:35:10 crc kubenswrapper[4980]: E1206 03:35:10.554370 4980 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 06 03:35:10 crc kubenswrapper[4980]: E1206 03:35:10.554498 4980 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/e6e88c5e-a255-449a-ae38-ac4d73a8e984-metrics-certs podName:e6e88c5e-a255-449a-ae38-ac4d73a8e984 nodeName:}" failed. No retries permitted until 2025-12-06 03:35:42.554471024 +0000 UTC m=+101.799222295 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/e6e88c5e-a255-449a-ae38-ac4d73a8e984-metrics-certs") pod "network-metrics-daemon-mcprh" (UID: "e6e88c5e-a255-449a-ae38-ac4d73a8e984") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 06 03:35:10 crc kubenswrapper[4980]: I1206 03:35:10.575753 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:10 crc kubenswrapper[4980]: I1206 03:35:10.575801 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:10 crc kubenswrapper[4980]: I1206 03:35:10.575812 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:10 crc kubenswrapper[4980]: I1206 03:35:10.575829 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:10 crc kubenswrapper[4980]: I1206 03:35:10.575841 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:10Z","lastTransitionTime":"2025-12-06T03:35:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:10 crc kubenswrapper[4980]: I1206 03:35:10.677996 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:10 crc kubenswrapper[4980]: I1206 03:35:10.678032 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:10 crc kubenswrapper[4980]: I1206 03:35:10.678041 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:10 crc kubenswrapper[4980]: I1206 03:35:10.678054 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:10 crc kubenswrapper[4980]: I1206 03:35:10.678064 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:10Z","lastTransitionTime":"2025-12-06T03:35:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:10 crc kubenswrapper[4980]: I1206 03:35:10.781331 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:10 crc kubenswrapper[4980]: I1206 03:35:10.781373 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:10 crc kubenswrapper[4980]: I1206 03:35:10.781386 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:10 crc kubenswrapper[4980]: I1206 03:35:10.781403 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:10 crc kubenswrapper[4980]: I1206 03:35:10.781419 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:10Z","lastTransitionTime":"2025-12-06T03:35:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:10 crc kubenswrapper[4980]: I1206 03:35:10.883800 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:10 crc kubenswrapper[4980]: I1206 03:35:10.883855 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:10 crc kubenswrapper[4980]: I1206 03:35:10.883866 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:10 crc kubenswrapper[4980]: I1206 03:35:10.883885 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:10 crc kubenswrapper[4980]: I1206 03:35:10.883900 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:10Z","lastTransitionTime":"2025-12-06T03:35:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:10 crc kubenswrapper[4980]: I1206 03:35:10.986361 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:10 crc kubenswrapper[4980]: I1206 03:35:10.986414 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:10 crc kubenswrapper[4980]: I1206 03:35:10.986429 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:10 crc kubenswrapper[4980]: I1206 03:35:10.986447 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:10 crc kubenswrapper[4980]: I1206 03:35:10.986459 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:10Z","lastTransitionTime":"2025-12-06T03:35:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:11 crc kubenswrapper[4980]: I1206 03:35:11.089887 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:11 crc kubenswrapper[4980]: I1206 03:35:11.089927 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:11 crc kubenswrapper[4980]: I1206 03:35:11.089938 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:11 crc kubenswrapper[4980]: I1206 03:35:11.089953 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:11 crc kubenswrapper[4980]: I1206 03:35:11.089961 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:11Z","lastTransitionTime":"2025-12-06T03:35:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:11 crc kubenswrapper[4980]: I1206 03:35:11.184488 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 03:35:11 crc kubenswrapper[4980]: E1206 03:35:11.184674 4980 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 03:35:11 crc kubenswrapper[4980]: I1206 03:35:11.192711 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:11 crc kubenswrapper[4980]: I1206 03:35:11.192777 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:11 crc kubenswrapper[4980]: I1206 03:35:11.192794 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:11 crc kubenswrapper[4980]: I1206 03:35:11.192812 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:11 crc kubenswrapper[4980]: I1206 03:35:11.192824 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:11Z","lastTransitionTime":"2025-12-06T03:35:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:11 crc kubenswrapper[4980]: I1206 03:35:11.196168 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://502c727d6baa31f400bee14c34b19abbe38af873f28f6a82a43547431755d981\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:35:11Z is after 2025-08-24T17:21:41Z" Dec 06 03:35:11 crc kubenswrapper[4980]: I1206 03:35:11.215538 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-dx44q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5bac22a-512c-49d0-90db-47de6aaaefd0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://97469dea2cd5393da0acf784ccc1969c8a4d07da7d51b14cdc955013ee9710ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f7mdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3b5dab476c7f227a6e4a2eae8469432667fb9fad345d9419ba59feab9fd385e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c3b5dab476c7f227a6e4a2eae8469432667fb9fad345d9419ba59feab9fd385e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f7mdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e921499231cbdf26b8918c98f9ad8c322fd7848865caeef048470de3e2c0efba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e921499231cbdf26b8918c98f9ad8c322fd7848865caeef048470de3e2c0efba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f7mdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7cbfe7d8cdb89916c238296592fab45b9e92e2c585baa6d67230decd1c84c5bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7cbfe7d8cdb89916c238296592fab45b9e92e2c585baa6d67230decd1c84c5bb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f7mdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f45cbf23b6b0d208cefc4b493f901bab9d71ccf2f1456507d54fb1f30c4fd48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7f45cbf23b6b0d208cefc4b493f901bab9d71ccf2f1456507d54fb1f30c4fd48\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f7mdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://344ce4b145ff021fcaf163803d8ad8cc6e9565bf7f0992e2653c43db663d8411\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://344ce4b145ff021fcaf163803d8ad8cc6e9565bf7f0992e2653c43db663d8411\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f7mdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://619e56a9937dd6376ab471ba16da30e5d77f443f5bd33610bea5ec85cd0446ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://619e56a9937dd6376ab471ba16da30e5d77f443f5bd33610bea5ec85cd0446ef\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f7mdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:23Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-dx44q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:35:11Z is after 2025-08-24T17:21:41Z" Dec 06 03:35:11 crc kubenswrapper[4980]: I1206 03:35:11.250208 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-r5zfb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"320f44d1-a671-4a91-b328-a8b0fdd8f23a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://92393ecaaef8cfcfd01837671f8cc89803dd570b637b4a373b40b7f487788826\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zhm5t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52e8db644463794041978ced2c0591c13e0ada8e5f1a78a94a0a665218a37b5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zhm5t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:23Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-r5zfb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:35:11Z is after 2025-08-24T17:21:41Z" Dec 06 03:35:11 crc kubenswrapper[4980]: I1206 03:35:11.259942 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-cvzg8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f1cde706-268d-4d26-818f-2f42bfb37b5f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://86f9a77acae360740be827eb0d1e1516e8a49cb9b197b6f5f9468ac70a101158\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rk86l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:29Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-cvzg8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:35:11Z is after 2025-08-24T17:21:41Z" Dec 06 03:35:11 crc kubenswrapper[4980]: I1206 03:35:11.277770 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cab608b3-a02f-475b-9aae-9b8f5d878de9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f7fcc44b42bcbb55509048e3a566bb557333c4a83e33d7b0fba8eb514317909\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c22e7dd8886c2dfbb742900ed86ec7b0c2edf9a1363e74f6666961d928f06c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://473e792b2d0a9644854ee0581c6120fc9c989bc342899ef72b1d056080fd4694\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ee8e936dfc39361d6f99460910591ca8c901a20f0d274bbb6aaebfa8a00ee06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://77bc12768b018625ba11a6bfd66688f926c4b7f3bc1d0e41c6551a18b485ecf3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://27cf07cd96893900da4afecd0d56fd9269bc60c1d23246b4685fcd8756357c19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://27cf07cd96893900da4afecd0d56fd9269bc60c1d23246b4685fcd8756357c19\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c49feb24566a8284378670fa0390710d38c237acc94ffb654f90f3fa4626e4c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c49feb24566a8284378670fa0390710d38c237acc94ffb654f90f3fa4626e4c8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:03Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://bb74d1c3d752a379fa394c450217badb45511fbcafad62f7b618131cab94b29e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bb74d1c3d752a379fa394c450217badb45511fbcafad62f7b618131cab94b29e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:01Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:35:11Z is after 2025-08-24T17:21:41Z" Dec 06 03:35:11 crc kubenswrapper[4980]: I1206 03:35:11.294583 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:11 crc kubenswrapper[4980]: I1206 03:35:11.294617 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:11 crc kubenswrapper[4980]: I1206 03:35:11.294625 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:11 crc kubenswrapper[4980]: I1206 03:35:11.294638 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:11 crc kubenswrapper[4980]: I1206 03:35:11.294648 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:11Z","lastTransitionTime":"2025-12-06T03:35:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:11 crc kubenswrapper[4980]: I1206 03:35:11.299173 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4ef01e7d9809fd9cfc46e3a58b07159cdc6f5a3c210d1c9efc49bca27b30858\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:35:11Z is after 2025-08-24T17:21:41Z" Dec 06 03:35:11 crc kubenswrapper[4980]: I1206 03:35:11.314791 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-kq4z8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f96c1f6c-55e8-4d86-bff3-5381581b7022\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://799498681afa7653408bc23b1e82d75c16e3c809d2ee7f322780e20c537eb710\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bdx7g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:23Z\\\"}}\" for pod \"openshift-multus\"/\"multus-kq4z8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:35:11Z is after 2025-08-24T17:21:41Z" Dec 06 03:35:11 crc kubenswrapper[4980]: I1206 03:35:11.325428 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-6kxfq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"951b6deb-bcea-4332-8ae8-c23ac7aaef36\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://423d30c3e65b04bdf6ff5ef53d8a7575a434c05e28489370a7370381ebc29de9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b8htf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://282df50dd9653c1cbe97fc2041d9e9ea91a992661682652aafc12e566d8f1b81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b8htf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:37Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-6kxfq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:35:11Z is after 2025-08-24T17:21:41Z" Dec 06 03:35:11 crc kubenswrapper[4980]: I1206 03:35:11.334752 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-mcprh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e6e88c5e-a255-449a-ae38-ac4d73a8e984\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g7xz9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g7xz9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:38Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-mcprh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:35:11Z is after 2025-08-24T17:21:41Z" Dec 06 03:35:11 crc kubenswrapper[4980]: I1206 03:35:11.345852 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:35:11Z is after 2025-08-24T17:21:41Z" Dec 06 03:35:11 crc kubenswrapper[4980]: I1206 03:35:11.356880 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7f958150-549d-4c80-8a7e-23c39e4f9500\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c28b6895b14e73f4f3eb7a8ce93076c05092ba9b58ea0e3e879c65bcfa227a21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://76cff8a2d652df38b04eefc053832baf4cbb5db778ff5e525b18c2825141b118\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://206f6f19fe3197f5d6798cc86c5473bf2decb02a2860ee7ffb5efb5f60952e8c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bafaa9670b05671fb57b658c2960bf3ee32c84e52f86a3a3bbab01d5a38ff0c4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://773a0bc2aab73ffc2717e987543defc9b786cc22300aa54b4ba4f3454504e331\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-06T03:34:22Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1206 03:34:15.305330 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1206 03:34:15.307249 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1592846546/tls.crt::/tmp/serving-cert-1592846546/tls.key\\\\\\\"\\\\nI1206 03:34:21.814037 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1206 03:34:21.816756 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1206 03:34:21.816780 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1206 03:34:21.816806 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1206 03:34:21.816813 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1206 03:34:21.831301 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1206 03:34:21.831311 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1206 03:34:21.831324 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 03:34:21.831330 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 03:34:21.831335 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1206 03:34:21.831339 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1206 03:34:21.831342 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1206 03:34:21.831346 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1206 03:34:21.911269 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:04Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d40e0e7446f8416bf9da94fc53c01be937f9a51a11eb101157d2ddd4597720d2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:04Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2cba89f2f1bed96d59f183628d900d9b72c91eb472e273c8bed0babcab485910\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2cba89f2f1bed96d59f183628d900d9b72c91eb472e273c8bed0babcab485910\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:01Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:35:11Z is after 2025-08-24T17:21:41Z" Dec 06 03:35:11 crc kubenswrapper[4980]: I1206 03:35:11.368407 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6ed68f66-1cf6-465a-8f66-76d5af6a44d8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://725e84f6bdfb722a8ce69158d03ea4c66a9f087642d55cb9651e003c81be0c8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://71f7b492ea07cdf7dc65fc48184ec1e64914eb1811bbfdddc0fb14c8ad6cb296\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b543aa4ee03d5a14909b67d876c5675c02e1df2a89a3c9ac8d3fb587fea53d8f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://32af3427e872a70cb3c2357116f04cabc8d3d9b69e7306b2e0de2ae22f9469d4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:01Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:35:11Z is after 2025-08-24T17:21:41Z" Dec 06 03:35:11 crc kubenswrapper[4980]: I1206 03:35:11.380340 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:22Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:35:11Z is after 2025-08-24T17:21:41Z" Dec 06 03:35:11 crc kubenswrapper[4980]: I1206 03:35:11.391552 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://334ea6a273fd7d268906c9f4a6dbf6f4df4fcdbf38f3d7d248cb7d6cf8537794\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ec44dcd5e4bf0a0466bb586ebd61259364dbbfe22499f4cbff854e3bd153082\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:35:11Z is after 2025-08-24T17:21:41Z" Dec 06 03:35:11 crc kubenswrapper[4980]: I1206 03:35:11.397179 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:11 crc kubenswrapper[4980]: I1206 03:35:11.397217 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:11 crc kubenswrapper[4980]: I1206 03:35:11.397225 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:11 crc kubenswrapper[4980]: I1206 03:35:11.397243 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:11 crc kubenswrapper[4980]: I1206 03:35:11.397255 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:11Z","lastTransitionTime":"2025-12-06T03:35:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:11 crc kubenswrapper[4980]: I1206 03:35:11.404302 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:22Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:35:11Z is after 2025-08-24T17:21:41Z" Dec 06 03:35:11 crc kubenswrapper[4980]: I1206 03:35:11.414919 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-sv5sz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"873977aa-4eb0-4526-bc0b-757ad2dea2a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01f0cc855c822a71ea8ef6a7ce6e25747ff9956691735973f935c4a28672a7a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lg7tx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:23Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-sv5sz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:35:11Z is after 2025-08-24T17:21:41Z" Dec 06 03:35:11 crc kubenswrapper[4980]: I1206 03:35:11.436041 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5w4fr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5af0e768-3c9c-4401-ab56-57516cd3170e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:23Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:23Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://768b8e1fe1a00a8467d744da5ad3f61006b0a3612c96f8e4d0ee6fcb93195f9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfzrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6342903b318317ee8d84c2f5d467cb25c811a789be0d6ab0e0cba37b5611e16d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfzrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d620685574db7594132ae24a91ec0a93dce20a9a289540deebb39650b8a1816\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfzrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c67ec8f9c1764fcdf215d52822c2d4f0437dc7f056f05d708767121f6c1f3515\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfzrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f35beae5339eeaf26494ed678e3c1bd02349eae2a4b4615325b17923f3b653b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfzrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a940ec6ee523313728f7e8ae934b7e0b0cff82771e3e64723309db1fae013c3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfzrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fae428d1a6eaeb2e6ddd4a5b1a1170e230635739ea0b39743e565e47fb7ee6e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fae428d1a6eaeb2e6ddd4a5b1a1170e230635739ea0b39743e565e47fb7ee6e4\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-06T03:34:53Z\\\",\\\"message\\\":\\\"/pkg/client/informers/externalversions/factory.go:141\\\\nI1206 03:34:53.908007 6569 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1206 03:34:53.908049 6569 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1206 03:34:53.908094 6569 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1206 03:34:53.908109 6569 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1206 03:34:53.908160 6569 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1206 03:34:53.908208 6569 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1206 03:34:53.908256 6569 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1206 03:34:53.908286 6569 factory.go:656] Stopping watch factory\\\\nI1206 03:34:53.908308 6569 ovnkube.go:599] Stopped ovnkube\\\\nI1206 03:34:53.908349 6569 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1206 03:34:53.908364 6569 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1206 03:34:53.908372 6569 handler.go:208] Removed *v1.Node event handler 2\\\\nI1206 03:34:53.908385 6569 handler.go:208] Removed *v1.Node event handler 7\\\\nI1206 03:34:53.908397 6569 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1206 03:34:53.908415 6569 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1206 03:34:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:52Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-5w4fr_openshift-ovn-kubernetes(5af0e768-3c9c-4401-ab56-57516cd3170e)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfzrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe261f4322b85702181d7473ab2a9a4d2d4b869d586c3091f811b12a81a63aae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfzrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d876df04b5ab0ba07a5cbcafbe4648403c5014c59eb7d40adee2657c3a255473\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d876df04b5ab0ba07a5cbcafbe4648403c5014c59eb7d40adee2657c3a255473\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfzrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:23Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-5w4fr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:35:11Z is after 2025-08-24T17:21:41Z" Dec 06 03:35:11 crc kubenswrapper[4980]: I1206 03:35:11.449972 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"64278250-0b19-4591-b06e-6c96652ee92a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5beac3fd93b448506c0a110d70ef5faf42efa18b16abe6b3d4eb7b9c4233ad93\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ae66052702fcfd94f0123f4fbbe9efed37eb04609269a2bc951f95669de16601\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://173903f3554d7ac9158493498a016d618d92f46da1500408834b3c24249cf117\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1f596bfd5321ca225f4be03d4cf229c31a52def9731c0e177ab3a0bc010380e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1f596bfd5321ca225f4be03d4cf229c31a52def9731c0e177ab3a0bc010380e1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:02Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:01Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:35:11Z is after 2025-08-24T17:21:41Z" Dec 06 03:35:11 crc kubenswrapper[4980]: I1206 03:35:11.499279 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:11 crc kubenswrapper[4980]: I1206 03:35:11.499314 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:11 crc kubenswrapper[4980]: I1206 03:35:11.499324 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:11 crc kubenswrapper[4980]: I1206 03:35:11.499340 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:11 crc kubenswrapper[4980]: I1206 03:35:11.499350 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:11Z","lastTransitionTime":"2025-12-06T03:35:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:11 crc kubenswrapper[4980]: I1206 03:35:11.601971 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:11 crc kubenswrapper[4980]: I1206 03:35:11.602116 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:11 crc kubenswrapper[4980]: I1206 03:35:11.602165 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:11 crc kubenswrapper[4980]: I1206 03:35:11.602187 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:11 crc kubenswrapper[4980]: I1206 03:35:11.602201 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:11Z","lastTransitionTime":"2025-12-06T03:35:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:11 crc kubenswrapper[4980]: I1206 03:35:11.705354 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:11 crc kubenswrapper[4980]: I1206 03:35:11.705731 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:11 crc kubenswrapper[4980]: I1206 03:35:11.705883 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:11 crc kubenswrapper[4980]: I1206 03:35:11.706027 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:11 crc kubenswrapper[4980]: I1206 03:35:11.706160 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:11Z","lastTransitionTime":"2025-12-06T03:35:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:11 crc kubenswrapper[4980]: I1206 03:35:11.809425 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:11 crc kubenswrapper[4980]: I1206 03:35:11.809473 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:11 crc kubenswrapper[4980]: I1206 03:35:11.809486 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:11 crc kubenswrapper[4980]: I1206 03:35:11.809502 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:11 crc kubenswrapper[4980]: I1206 03:35:11.809534 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:11Z","lastTransitionTime":"2025-12-06T03:35:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:11 crc kubenswrapper[4980]: I1206 03:35:11.912310 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:11 crc kubenswrapper[4980]: I1206 03:35:11.912360 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:11 crc kubenswrapper[4980]: I1206 03:35:11.912376 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:11 crc kubenswrapper[4980]: I1206 03:35:11.912729 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:11 crc kubenswrapper[4980]: I1206 03:35:11.912749 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:11Z","lastTransitionTime":"2025-12-06T03:35:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:12 crc kubenswrapper[4980]: I1206 03:35:12.016277 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:12 crc kubenswrapper[4980]: I1206 03:35:12.016335 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:12 crc kubenswrapper[4980]: I1206 03:35:12.016365 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:12 crc kubenswrapper[4980]: I1206 03:35:12.016388 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:12 crc kubenswrapper[4980]: I1206 03:35:12.016408 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:12Z","lastTransitionTime":"2025-12-06T03:35:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:12 crc kubenswrapper[4980]: I1206 03:35:12.119815 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:12 crc kubenswrapper[4980]: I1206 03:35:12.120166 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:12 crc kubenswrapper[4980]: I1206 03:35:12.120299 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:12 crc kubenswrapper[4980]: I1206 03:35:12.120447 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:12 crc kubenswrapper[4980]: I1206 03:35:12.120598 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:12Z","lastTransitionTime":"2025-12-06T03:35:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:12 crc kubenswrapper[4980]: I1206 03:35:12.183352 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-mcprh" Dec 06 03:35:12 crc kubenswrapper[4980]: I1206 03:35:12.183449 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 03:35:12 crc kubenswrapper[4980]: I1206 03:35:12.183449 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 03:35:12 crc kubenswrapper[4980]: E1206 03:35:12.183889 4980 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 03:35:12 crc kubenswrapper[4980]: E1206 03:35:12.183992 4980 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 03:35:12 crc kubenswrapper[4980]: E1206 03:35:12.183743 4980 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-mcprh" podUID="e6e88c5e-a255-449a-ae38-ac4d73a8e984" Dec 06 03:35:12 crc kubenswrapper[4980]: I1206 03:35:12.223577 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:12 crc kubenswrapper[4980]: I1206 03:35:12.223635 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:12 crc kubenswrapper[4980]: I1206 03:35:12.223651 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:12 crc kubenswrapper[4980]: I1206 03:35:12.223678 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:12 crc kubenswrapper[4980]: I1206 03:35:12.223693 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:12Z","lastTransitionTime":"2025-12-06T03:35:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:12 crc kubenswrapper[4980]: I1206 03:35:12.327910 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:12 crc kubenswrapper[4980]: I1206 03:35:12.327968 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:12 crc kubenswrapper[4980]: I1206 03:35:12.327981 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:12 crc kubenswrapper[4980]: I1206 03:35:12.327996 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:12 crc kubenswrapper[4980]: I1206 03:35:12.328008 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:12Z","lastTransitionTime":"2025-12-06T03:35:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:12 crc kubenswrapper[4980]: I1206 03:35:12.431536 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:12 crc kubenswrapper[4980]: I1206 03:35:12.431735 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:12 crc kubenswrapper[4980]: I1206 03:35:12.431766 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:12 crc kubenswrapper[4980]: I1206 03:35:12.431796 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:12 crc kubenswrapper[4980]: I1206 03:35:12.431820 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:12Z","lastTransitionTime":"2025-12-06T03:35:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:12 crc kubenswrapper[4980]: I1206 03:35:12.534464 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:12 crc kubenswrapper[4980]: I1206 03:35:12.534542 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:12 crc kubenswrapper[4980]: I1206 03:35:12.534553 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:12 crc kubenswrapper[4980]: I1206 03:35:12.534568 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:12 crc kubenswrapper[4980]: I1206 03:35:12.534578 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:12Z","lastTransitionTime":"2025-12-06T03:35:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:12 crc kubenswrapper[4980]: I1206 03:35:12.637842 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:12 crc kubenswrapper[4980]: I1206 03:35:12.637892 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:12 crc kubenswrapper[4980]: I1206 03:35:12.637904 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:12 crc kubenswrapper[4980]: I1206 03:35:12.637922 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:12 crc kubenswrapper[4980]: I1206 03:35:12.637934 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:12Z","lastTransitionTime":"2025-12-06T03:35:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:12 crc kubenswrapper[4980]: I1206 03:35:12.740564 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:12 crc kubenswrapper[4980]: I1206 03:35:12.740622 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:12 crc kubenswrapper[4980]: I1206 03:35:12.740635 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:12 crc kubenswrapper[4980]: I1206 03:35:12.740655 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:12 crc kubenswrapper[4980]: I1206 03:35:12.740667 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:12Z","lastTransitionTime":"2025-12-06T03:35:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:12 crc kubenswrapper[4980]: I1206 03:35:12.843310 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:12 crc kubenswrapper[4980]: I1206 03:35:12.843380 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:12 crc kubenswrapper[4980]: I1206 03:35:12.843398 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:12 crc kubenswrapper[4980]: I1206 03:35:12.843424 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:12 crc kubenswrapper[4980]: I1206 03:35:12.843442 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:12Z","lastTransitionTime":"2025-12-06T03:35:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:12 crc kubenswrapper[4980]: I1206 03:35:12.961268 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:12 crc kubenswrapper[4980]: I1206 03:35:12.961401 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:12 crc kubenswrapper[4980]: I1206 03:35:12.961420 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:12 crc kubenswrapper[4980]: I1206 03:35:12.961477 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:12 crc kubenswrapper[4980]: I1206 03:35:12.961498 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:12Z","lastTransitionTime":"2025-12-06T03:35:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:13 crc kubenswrapper[4980]: I1206 03:35:13.064842 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:13 crc kubenswrapper[4980]: I1206 03:35:13.064900 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:13 crc kubenswrapper[4980]: I1206 03:35:13.064912 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:13 crc kubenswrapper[4980]: I1206 03:35:13.064930 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:13 crc kubenswrapper[4980]: I1206 03:35:13.064941 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:13Z","lastTransitionTime":"2025-12-06T03:35:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:13 crc kubenswrapper[4980]: I1206 03:35:13.168628 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:13 crc kubenswrapper[4980]: I1206 03:35:13.168698 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:13 crc kubenswrapper[4980]: I1206 03:35:13.168719 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:13 crc kubenswrapper[4980]: I1206 03:35:13.168767 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:13 crc kubenswrapper[4980]: I1206 03:35:13.168795 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:13Z","lastTransitionTime":"2025-12-06T03:35:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:13 crc kubenswrapper[4980]: I1206 03:35:13.184068 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 03:35:13 crc kubenswrapper[4980]: E1206 03:35:13.184327 4980 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 03:35:13 crc kubenswrapper[4980]: I1206 03:35:13.271495 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:13 crc kubenswrapper[4980]: I1206 03:35:13.272251 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:13 crc kubenswrapper[4980]: I1206 03:35:13.272309 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:13 crc kubenswrapper[4980]: I1206 03:35:13.272346 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:13 crc kubenswrapper[4980]: I1206 03:35:13.272372 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:13Z","lastTransitionTime":"2025-12-06T03:35:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:13 crc kubenswrapper[4980]: I1206 03:35:13.374985 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:13 crc kubenswrapper[4980]: I1206 03:35:13.375045 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:13 crc kubenswrapper[4980]: I1206 03:35:13.375062 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:13 crc kubenswrapper[4980]: I1206 03:35:13.375088 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:13 crc kubenswrapper[4980]: I1206 03:35:13.375103 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:13Z","lastTransitionTime":"2025-12-06T03:35:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:13 crc kubenswrapper[4980]: I1206 03:35:13.478363 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:13 crc kubenswrapper[4980]: I1206 03:35:13.478439 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:13 crc kubenswrapper[4980]: I1206 03:35:13.478462 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:13 crc kubenswrapper[4980]: I1206 03:35:13.478494 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:13 crc kubenswrapper[4980]: I1206 03:35:13.478546 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:13Z","lastTransitionTime":"2025-12-06T03:35:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:13 crc kubenswrapper[4980]: I1206 03:35:13.581326 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:13 crc kubenswrapper[4980]: I1206 03:35:13.581365 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:13 crc kubenswrapper[4980]: I1206 03:35:13.581377 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:13 crc kubenswrapper[4980]: I1206 03:35:13.581396 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:13 crc kubenswrapper[4980]: I1206 03:35:13.581408 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:13Z","lastTransitionTime":"2025-12-06T03:35:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:13 crc kubenswrapper[4980]: I1206 03:35:13.684092 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:13 crc kubenswrapper[4980]: I1206 03:35:13.684165 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:13 crc kubenswrapper[4980]: I1206 03:35:13.684199 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:13 crc kubenswrapper[4980]: I1206 03:35:13.684228 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:13 crc kubenswrapper[4980]: I1206 03:35:13.684249 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:13Z","lastTransitionTime":"2025-12-06T03:35:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:13 crc kubenswrapper[4980]: I1206 03:35:13.787308 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:13 crc kubenswrapper[4980]: I1206 03:35:13.787342 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:13 crc kubenswrapper[4980]: I1206 03:35:13.787354 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:13 crc kubenswrapper[4980]: I1206 03:35:13.787371 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:13 crc kubenswrapper[4980]: I1206 03:35:13.787382 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:13Z","lastTransitionTime":"2025-12-06T03:35:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:13 crc kubenswrapper[4980]: I1206 03:35:13.890506 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:13 crc kubenswrapper[4980]: I1206 03:35:13.890626 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:13 crc kubenswrapper[4980]: I1206 03:35:13.890655 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:13 crc kubenswrapper[4980]: I1206 03:35:13.890742 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:13 crc kubenswrapper[4980]: I1206 03:35:13.890762 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:13Z","lastTransitionTime":"2025-12-06T03:35:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:13 crc kubenswrapper[4980]: I1206 03:35:13.994169 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:13 crc kubenswrapper[4980]: I1206 03:35:13.994213 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:13 crc kubenswrapper[4980]: I1206 03:35:13.994226 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:13 crc kubenswrapper[4980]: I1206 03:35:13.994243 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:13 crc kubenswrapper[4980]: I1206 03:35:13.994255 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:13Z","lastTransitionTime":"2025-12-06T03:35:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:14 crc kubenswrapper[4980]: I1206 03:35:14.096787 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:14 crc kubenswrapper[4980]: I1206 03:35:14.096824 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:14 crc kubenswrapper[4980]: I1206 03:35:14.096834 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:14 crc kubenswrapper[4980]: I1206 03:35:14.096849 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:14 crc kubenswrapper[4980]: I1206 03:35:14.096860 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:14Z","lastTransitionTime":"2025-12-06T03:35:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:14 crc kubenswrapper[4980]: I1206 03:35:14.183718 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 03:35:14 crc kubenswrapper[4980]: I1206 03:35:14.183883 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-mcprh" Dec 06 03:35:14 crc kubenswrapper[4980]: E1206 03:35:14.184085 4980 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 03:35:14 crc kubenswrapper[4980]: I1206 03:35:14.184205 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 03:35:14 crc kubenswrapper[4980]: E1206 03:35:14.184350 4980 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-mcprh" podUID="e6e88c5e-a255-449a-ae38-ac4d73a8e984" Dec 06 03:35:14 crc kubenswrapper[4980]: E1206 03:35:14.184555 4980 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 03:35:14 crc kubenswrapper[4980]: I1206 03:35:14.199381 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:14 crc kubenswrapper[4980]: I1206 03:35:14.199415 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:14 crc kubenswrapper[4980]: I1206 03:35:14.199423 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:14 crc kubenswrapper[4980]: I1206 03:35:14.199436 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:14 crc kubenswrapper[4980]: I1206 03:35:14.199446 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:14Z","lastTransitionTime":"2025-12-06T03:35:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:14 crc kubenswrapper[4980]: I1206 03:35:14.302586 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:14 crc kubenswrapper[4980]: I1206 03:35:14.302643 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:14 crc kubenswrapper[4980]: I1206 03:35:14.302655 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:14 crc kubenswrapper[4980]: I1206 03:35:14.302673 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:14 crc kubenswrapper[4980]: I1206 03:35:14.302685 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:14Z","lastTransitionTime":"2025-12-06T03:35:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:14 crc kubenswrapper[4980]: I1206 03:35:14.405344 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:14 crc kubenswrapper[4980]: I1206 03:35:14.405375 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:14 crc kubenswrapper[4980]: I1206 03:35:14.405384 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:14 crc kubenswrapper[4980]: I1206 03:35:14.405397 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:14 crc kubenswrapper[4980]: I1206 03:35:14.405405 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:14Z","lastTransitionTime":"2025-12-06T03:35:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:14 crc kubenswrapper[4980]: I1206 03:35:14.508421 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:14 crc kubenswrapper[4980]: I1206 03:35:14.508463 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:14 crc kubenswrapper[4980]: I1206 03:35:14.508481 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:14 crc kubenswrapper[4980]: I1206 03:35:14.508503 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:14 crc kubenswrapper[4980]: I1206 03:35:14.508550 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:14Z","lastTransitionTime":"2025-12-06T03:35:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:14 crc kubenswrapper[4980]: I1206 03:35:14.612432 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:14 crc kubenswrapper[4980]: I1206 03:35:14.612505 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:14 crc kubenswrapper[4980]: I1206 03:35:14.612557 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:14 crc kubenswrapper[4980]: I1206 03:35:14.612581 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:14 crc kubenswrapper[4980]: I1206 03:35:14.612600 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:14Z","lastTransitionTime":"2025-12-06T03:35:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:14 crc kubenswrapper[4980]: I1206 03:35:14.717045 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:14 crc kubenswrapper[4980]: I1206 03:35:14.717097 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:14 crc kubenswrapper[4980]: I1206 03:35:14.717121 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:14 crc kubenswrapper[4980]: I1206 03:35:14.717148 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:14 crc kubenswrapper[4980]: I1206 03:35:14.717169 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:14Z","lastTransitionTime":"2025-12-06T03:35:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:14 crc kubenswrapper[4980]: I1206 03:35:14.820824 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:14 crc kubenswrapper[4980]: I1206 03:35:14.820891 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:14 crc kubenswrapper[4980]: I1206 03:35:14.820912 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:14 crc kubenswrapper[4980]: I1206 03:35:14.820938 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:14 crc kubenswrapper[4980]: I1206 03:35:14.820956 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:14Z","lastTransitionTime":"2025-12-06T03:35:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:14 crc kubenswrapper[4980]: I1206 03:35:14.923805 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:14 crc kubenswrapper[4980]: I1206 03:35:14.923884 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:14 crc kubenswrapper[4980]: I1206 03:35:14.923907 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:14 crc kubenswrapper[4980]: I1206 03:35:14.923937 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:14 crc kubenswrapper[4980]: I1206 03:35:14.923957 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:14Z","lastTransitionTime":"2025-12-06T03:35:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:15 crc kubenswrapper[4980]: I1206 03:35:15.027159 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:15 crc kubenswrapper[4980]: I1206 03:35:15.027229 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:15 crc kubenswrapper[4980]: I1206 03:35:15.027248 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:15 crc kubenswrapper[4980]: I1206 03:35:15.027270 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:15 crc kubenswrapper[4980]: I1206 03:35:15.027287 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:15Z","lastTransitionTime":"2025-12-06T03:35:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:15 crc kubenswrapper[4980]: I1206 03:35:15.130580 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:15 crc kubenswrapper[4980]: I1206 03:35:15.130636 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:15 crc kubenswrapper[4980]: I1206 03:35:15.130682 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:15 crc kubenswrapper[4980]: I1206 03:35:15.130702 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:15 crc kubenswrapper[4980]: I1206 03:35:15.130715 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:15Z","lastTransitionTime":"2025-12-06T03:35:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:15 crc kubenswrapper[4980]: I1206 03:35:15.184159 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 03:35:15 crc kubenswrapper[4980]: E1206 03:35:15.184400 4980 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 03:35:15 crc kubenswrapper[4980]: I1206 03:35:15.234007 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:15 crc kubenswrapper[4980]: I1206 03:35:15.234078 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:15 crc kubenswrapper[4980]: I1206 03:35:15.234100 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:15 crc kubenswrapper[4980]: I1206 03:35:15.234132 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:15 crc kubenswrapper[4980]: I1206 03:35:15.234154 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:15Z","lastTransitionTime":"2025-12-06T03:35:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:15 crc kubenswrapper[4980]: I1206 03:35:15.338151 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:15 crc kubenswrapper[4980]: I1206 03:35:15.338229 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:15 crc kubenswrapper[4980]: I1206 03:35:15.338256 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:15 crc kubenswrapper[4980]: I1206 03:35:15.338290 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:15 crc kubenswrapper[4980]: I1206 03:35:15.338311 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:15Z","lastTransitionTime":"2025-12-06T03:35:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:15 crc kubenswrapper[4980]: I1206 03:35:15.444108 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:15 crc kubenswrapper[4980]: I1206 03:35:15.444226 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:15 crc kubenswrapper[4980]: I1206 03:35:15.444257 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:15 crc kubenswrapper[4980]: I1206 03:35:15.444287 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:15 crc kubenswrapper[4980]: I1206 03:35:15.444309 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:15Z","lastTransitionTime":"2025-12-06T03:35:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:15 crc kubenswrapper[4980]: I1206 03:35:15.548199 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:15 crc kubenswrapper[4980]: I1206 03:35:15.548264 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:15 crc kubenswrapper[4980]: I1206 03:35:15.548284 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:15 crc kubenswrapper[4980]: I1206 03:35:15.548312 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:15 crc kubenswrapper[4980]: I1206 03:35:15.548351 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:15Z","lastTransitionTime":"2025-12-06T03:35:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:15 crc kubenswrapper[4980]: I1206 03:35:15.651085 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:15 crc kubenswrapper[4980]: I1206 03:35:15.651142 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:15 crc kubenswrapper[4980]: I1206 03:35:15.651161 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:15 crc kubenswrapper[4980]: I1206 03:35:15.651198 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:15 crc kubenswrapper[4980]: I1206 03:35:15.651238 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:15Z","lastTransitionTime":"2025-12-06T03:35:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:15 crc kubenswrapper[4980]: I1206 03:35:15.754314 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:15 crc kubenswrapper[4980]: I1206 03:35:15.754360 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:15 crc kubenswrapper[4980]: I1206 03:35:15.754372 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:15 crc kubenswrapper[4980]: I1206 03:35:15.754389 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:15 crc kubenswrapper[4980]: I1206 03:35:15.754400 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:15Z","lastTransitionTime":"2025-12-06T03:35:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:15 crc kubenswrapper[4980]: I1206 03:35:15.857804 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:15 crc kubenswrapper[4980]: I1206 03:35:15.857884 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:15 crc kubenswrapper[4980]: I1206 03:35:15.857907 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:15 crc kubenswrapper[4980]: I1206 03:35:15.857937 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:15 crc kubenswrapper[4980]: I1206 03:35:15.857962 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:15Z","lastTransitionTime":"2025-12-06T03:35:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:15 crc kubenswrapper[4980]: I1206 03:35:15.961140 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:15 crc kubenswrapper[4980]: I1206 03:35:15.961219 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:15 crc kubenswrapper[4980]: I1206 03:35:15.961243 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:15 crc kubenswrapper[4980]: I1206 03:35:15.961272 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:15 crc kubenswrapper[4980]: I1206 03:35:15.961294 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:15Z","lastTransitionTime":"2025-12-06T03:35:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:16 crc kubenswrapper[4980]: I1206 03:35:16.065068 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:16 crc kubenswrapper[4980]: I1206 03:35:16.065138 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:16 crc kubenswrapper[4980]: I1206 03:35:16.065201 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:16 crc kubenswrapper[4980]: I1206 03:35:16.065230 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:16 crc kubenswrapper[4980]: I1206 03:35:16.065250 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:16Z","lastTransitionTime":"2025-12-06T03:35:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:16 crc kubenswrapper[4980]: I1206 03:35:16.112971 4980 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-kq4z8_f96c1f6c-55e8-4d86-bff3-5381581b7022/kube-multus/0.log" Dec 06 03:35:16 crc kubenswrapper[4980]: I1206 03:35:16.113061 4980 generic.go:334] "Generic (PLEG): container finished" podID="f96c1f6c-55e8-4d86-bff3-5381581b7022" containerID="799498681afa7653408bc23b1e82d75c16e3c809d2ee7f322780e20c537eb710" exitCode=1 Dec 06 03:35:16 crc kubenswrapper[4980]: I1206 03:35:16.113146 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-kq4z8" event={"ID":"f96c1f6c-55e8-4d86-bff3-5381581b7022","Type":"ContainerDied","Data":"799498681afa7653408bc23b1e82d75c16e3c809d2ee7f322780e20c537eb710"} Dec 06 03:35:16 crc kubenswrapper[4980]: I1206 03:35:16.114088 4980 scope.go:117] "RemoveContainer" containerID="799498681afa7653408bc23b1e82d75c16e3c809d2ee7f322780e20c537eb710" Dec 06 03:35:16 crc kubenswrapper[4980]: I1206 03:35:16.154312 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6ed68f66-1cf6-465a-8f66-76d5af6a44d8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://725e84f6bdfb722a8ce69158d03ea4c66a9f087642d55cb9651e003c81be0c8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://71f7b492ea07cdf7dc65fc48184ec1e64914eb1811bbfdddc0fb14c8ad6cb296\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b543aa4ee03d5a14909b67d876c5675c02e1df2a89a3c9ac8d3fb587fea53d8f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://32af3427e872a70cb3c2357116f04cabc8d3d9b69e7306b2e0de2ae22f9469d4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:01Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:35:16Z is after 2025-08-24T17:21:41Z" Dec 06 03:35:16 crc kubenswrapper[4980]: I1206 03:35:16.172871 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:16 crc kubenswrapper[4980]: I1206 03:35:16.172956 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:16 crc kubenswrapper[4980]: I1206 03:35:16.172988 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:16 crc kubenswrapper[4980]: I1206 03:35:16.173014 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:16 crc kubenswrapper[4980]: I1206 03:35:16.173032 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:16Z","lastTransitionTime":"2025-12-06T03:35:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:16 crc kubenswrapper[4980]: I1206 03:35:16.177701 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7f958150-549d-4c80-8a7e-23c39e4f9500\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c28b6895b14e73f4f3eb7a8ce93076c05092ba9b58ea0e3e879c65bcfa227a21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://76cff8a2d652df38b04eefc053832baf4cbb5db778ff5e525b18c2825141b118\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://206f6f19fe3197f5d6798cc86c5473bf2decb02a2860ee7ffb5efb5f60952e8c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bafaa9670b05671fb57b658c2960bf3ee32c84e52f86a3a3bbab01d5a38ff0c4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://773a0bc2aab73ffc2717e987543defc9b786cc22300aa54b4ba4f3454504e331\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-06T03:34:22Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1206 03:34:15.305330 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1206 03:34:15.307249 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1592846546/tls.crt::/tmp/serving-cert-1592846546/tls.key\\\\\\\"\\\\nI1206 03:34:21.814037 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1206 03:34:21.816756 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1206 03:34:21.816780 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1206 03:34:21.816806 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1206 03:34:21.816813 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1206 03:34:21.831301 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1206 03:34:21.831311 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1206 03:34:21.831324 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 03:34:21.831330 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 03:34:21.831335 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1206 03:34:21.831339 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1206 03:34:21.831342 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1206 03:34:21.831346 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1206 03:34:21.911269 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:04Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d40e0e7446f8416bf9da94fc53c01be937f9a51a11eb101157d2ddd4597720d2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:04Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2cba89f2f1bed96d59f183628d900d9b72c91eb472e273c8bed0babcab485910\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2cba89f2f1bed96d59f183628d900d9b72c91eb472e273c8bed0babcab485910\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:01Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:35:16Z is after 2025-08-24T17:21:41Z" Dec 06 03:35:16 crc kubenswrapper[4980]: I1206 03:35:16.183691 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 03:35:16 crc kubenswrapper[4980]: E1206 03:35:16.183900 4980 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 03:35:16 crc kubenswrapper[4980]: I1206 03:35:16.184229 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-mcprh" Dec 06 03:35:16 crc kubenswrapper[4980]: I1206 03:35:16.184236 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 03:35:16 crc kubenswrapper[4980]: E1206 03:35:16.184376 4980 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-mcprh" podUID="e6e88c5e-a255-449a-ae38-ac4d73a8e984" Dec 06 03:35:16 crc kubenswrapper[4980]: E1206 03:35:16.184582 4980 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 03:35:16 crc kubenswrapper[4980]: I1206 03:35:16.193287 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://334ea6a273fd7d268906c9f4a6dbf6f4df4fcdbf38f3d7d248cb7d6cf8537794\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ec44dcd5e4bf0a0466bb586ebd61259364dbbfe22499f4cbff854e3bd153082\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:35:16Z is after 2025-08-24T17:21:41Z" Dec 06 03:35:16 crc kubenswrapper[4980]: I1206 03:35:16.210808 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:22Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:35:16Z is after 2025-08-24T17:21:41Z" Dec 06 03:35:16 crc kubenswrapper[4980]: I1206 03:35:16.222660 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-sv5sz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"873977aa-4eb0-4526-bc0b-757ad2dea2a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01f0cc855c822a71ea8ef6a7ce6e25747ff9956691735973f935c4a28672a7a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lg7tx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:23Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-sv5sz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:35:16Z is after 2025-08-24T17:21:41Z" Dec 06 03:35:16 crc kubenswrapper[4980]: I1206 03:35:16.247970 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5w4fr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5af0e768-3c9c-4401-ab56-57516cd3170e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:23Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:23Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://768b8e1fe1a00a8467d744da5ad3f61006b0a3612c96f8e4d0ee6fcb93195f9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfzrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6342903b318317ee8d84c2f5d467cb25c811a789be0d6ab0e0cba37b5611e16d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfzrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d620685574db7594132ae24a91ec0a93dce20a9a289540deebb39650b8a1816\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfzrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c67ec8f9c1764fcdf215d52822c2d4f0437dc7f056f05d708767121f6c1f3515\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfzrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f35beae5339eeaf26494ed678e3c1bd02349eae2a4b4615325b17923f3b653b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfzrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a940ec6ee523313728f7e8ae934b7e0b0cff82771e3e64723309db1fae013c3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfzrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fae428d1a6eaeb2e6ddd4a5b1a1170e230635739ea0b39743e565e47fb7ee6e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fae428d1a6eaeb2e6ddd4a5b1a1170e230635739ea0b39743e565e47fb7ee6e4\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-06T03:34:53Z\\\",\\\"message\\\":\\\"/pkg/client/informers/externalversions/factory.go:141\\\\nI1206 03:34:53.908007 6569 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1206 03:34:53.908049 6569 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1206 03:34:53.908094 6569 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1206 03:34:53.908109 6569 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1206 03:34:53.908160 6569 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1206 03:34:53.908208 6569 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1206 03:34:53.908256 6569 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1206 03:34:53.908286 6569 factory.go:656] Stopping watch factory\\\\nI1206 03:34:53.908308 6569 ovnkube.go:599] Stopped ovnkube\\\\nI1206 03:34:53.908349 6569 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1206 03:34:53.908364 6569 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1206 03:34:53.908372 6569 handler.go:208] Removed *v1.Node event handler 2\\\\nI1206 03:34:53.908385 6569 handler.go:208] Removed *v1.Node event handler 7\\\\nI1206 03:34:53.908397 6569 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1206 03:34:53.908415 6569 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1206 03:34:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:52Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-5w4fr_openshift-ovn-kubernetes(5af0e768-3c9c-4401-ab56-57516cd3170e)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfzrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe261f4322b85702181d7473ab2a9a4d2d4b869d586c3091f811b12a81a63aae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfzrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d876df04b5ab0ba07a5cbcafbe4648403c5014c59eb7d40adee2657c3a255473\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d876df04b5ab0ba07a5cbcafbe4648403c5014c59eb7d40adee2657c3a255473\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfzrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:23Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-5w4fr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:35:16Z is after 2025-08-24T17:21:41Z" Dec 06 03:35:16 crc kubenswrapper[4980]: I1206 03:35:16.262099 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"64278250-0b19-4591-b06e-6c96652ee92a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5beac3fd93b448506c0a110d70ef5faf42efa18b16abe6b3d4eb7b9c4233ad93\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ae66052702fcfd94f0123f4fbbe9efed37eb04609269a2bc951f95669de16601\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://173903f3554d7ac9158493498a016d618d92f46da1500408834b3c24249cf117\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1f596bfd5321ca225f4be03d4cf229c31a52def9731c0e177ab3a0bc010380e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1f596bfd5321ca225f4be03d4cf229c31a52def9731c0e177ab3a0bc010380e1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:02Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:01Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:35:16Z is after 2025-08-24T17:21:41Z" Dec 06 03:35:16 crc kubenswrapper[4980]: I1206 03:35:16.275832 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:22Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:35:16Z is after 2025-08-24T17:21:41Z" Dec 06 03:35:16 crc kubenswrapper[4980]: I1206 03:35:16.278743 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:16 crc kubenswrapper[4980]: I1206 03:35:16.278771 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:16 crc kubenswrapper[4980]: I1206 03:35:16.278784 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:16 crc kubenswrapper[4980]: I1206 03:35:16.278799 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:16 crc kubenswrapper[4980]: I1206 03:35:16.278807 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:16Z","lastTransitionTime":"2025-12-06T03:35:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:16 crc kubenswrapper[4980]: I1206 03:35:16.294772 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-dx44q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5bac22a-512c-49d0-90db-47de6aaaefd0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://97469dea2cd5393da0acf784ccc1969c8a4d07da7d51b14cdc955013ee9710ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f7mdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3b5dab476c7f227a6e4a2eae8469432667fb9fad345d9419ba59feab9fd385e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c3b5dab476c7f227a6e4a2eae8469432667fb9fad345d9419ba59feab9fd385e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f7mdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e921499231cbdf26b8918c98f9ad8c322fd7848865caeef048470de3e2c0efba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e921499231cbdf26b8918c98f9ad8c322fd7848865caeef048470de3e2c0efba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f7mdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7cbfe7d8cdb89916c238296592fab45b9e92e2c585baa6d67230decd1c84c5bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7cbfe7d8cdb89916c238296592fab45b9e92e2c585baa6d67230decd1c84c5bb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f7mdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f45cbf23b6b0d208cefc4b493f901bab9d71ccf2f1456507d54fb1f30c4fd48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7f45cbf23b6b0d208cefc4b493f901bab9d71ccf2f1456507d54fb1f30c4fd48\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f7mdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://344ce4b145ff021fcaf163803d8ad8cc6e9565bf7f0992e2653c43db663d8411\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://344ce4b145ff021fcaf163803d8ad8cc6e9565bf7f0992e2653c43db663d8411\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f7mdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://619e56a9937dd6376ab471ba16da30e5d77f443f5bd33610bea5ec85cd0446ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://619e56a9937dd6376ab471ba16da30e5d77f443f5bd33610bea5ec85cd0446ef\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f7mdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:23Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-dx44q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:35:16Z is after 2025-08-24T17:21:41Z" Dec 06 03:35:16 crc kubenswrapper[4980]: I1206 03:35:16.311754 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-r5zfb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"320f44d1-a671-4a91-b328-a8b0fdd8f23a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://92393ecaaef8cfcfd01837671f8cc89803dd570b637b4a373b40b7f487788826\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zhm5t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52e8db644463794041978ced2c0591c13e0ada8e5f1a78a94a0a665218a37b5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zhm5t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:23Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-r5zfb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:35:16Z is after 2025-08-24T17:21:41Z" Dec 06 03:35:16 crc kubenswrapper[4980]: I1206 03:35:16.322725 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-cvzg8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f1cde706-268d-4d26-818f-2f42bfb37b5f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://86f9a77acae360740be827eb0d1e1516e8a49cb9b197b6f5f9468ac70a101158\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rk86l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:29Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-cvzg8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:35:16Z is after 2025-08-24T17:21:41Z" Dec 06 03:35:16 crc kubenswrapper[4980]: I1206 03:35:16.347615 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cab608b3-a02f-475b-9aae-9b8f5d878de9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f7fcc44b42bcbb55509048e3a566bb557333c4a83e33d7b0fba8eb514317909\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c22e7dd8886c2dfbb742900ed86ec7b0c2edf9a1363e74f6666961d928f06c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://473e792b2d0a9644854ee0581c6120fc9c989bc342899ef72b1d056080fd4694\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ee8e936dfc39361d6f99460910591ca8c901a20f0d274bbb6aaebfa8a00ee06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://77bc12768b018625ba11a6bfd66688f926c4b7f3bc1d0e41c6551a18b485ecf3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://27cf07cd96893900da4afecd0d56fd9269bc60c1d23246b4685fcd8756357c19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://27cf07cd96893900da4afecd0d56fd9269bc60c1d23246b4685fcd8756357c19\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c49feb24566a8284378670fa0390710d38c237acc94ffb654f90f3fa4626e4c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c49feb24566a8284378670fa0390710d38c237acc94ffb654f90f3fa4626e4c8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:03Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://bb74d1c3d752a379fa394c450217badb45511fbcafad62f7b618131cab94b29e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bb74d1c3d752a379fa394c450217badb45511fbcafad62f7b618131cab94b29e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:01Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:35:16Z is after 2025-08-24T17:21:41Z" Dec 06 03:35:16 crc kubenswrapper[4980]: I1206 03:35:16.364449 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://502c727d6baa31f400bee14c34b19abbe38af873f28f6a82a43547431755d981\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:35:16Z is after 2025-08-24T17:21:41Z" Dec 06 03:35:16 crc kubenswrapper[4980]: I1206 03:35:16.381736 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:16 crc kubenswrapper[4980]: I1206 03:35:16.381784 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:16 crc kubenswrapper[4980]: I1206 03:35:16.381796 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:16 crc kubenswrapper[4980]: I1206 03:35:16.381815 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:16 crc kubenswrapper[4980]: I1206 03:35:16.381827 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:16Z","lastTransitionTime":"2025-12-06T03:35:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:16 crc kubenswrapper[4980]: I1206 03:35:16.384093 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-kq4z8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f96c1f6c-55e8-4d86-bff3-5381581b7022\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:35:16Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:35:16Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://799498681afa7653408bc23b1e82d75c16e3c809d2ee7f322780e20c537eb710\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://799498681afa7653408bc23b1e82d75c16e3c809d2ee7f322780e20c537eb710\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-06T03:35:15Z\\\",\\\"message\\\":\\\"2025-12-06T03:34:30+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_68ae7793-ff25-4e9c-8158-bea74df4ae16\\\\n2025-12-06T03:34:30+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_68ae7793-ff25-4e9c-8158-bea74df4ae16 to /host/opt/cni/bin/\\\\n2025-12-06T03:34:30Z [verbose] multus-daemon started\\\\n2025-12-06T03:34:30Z [verbose] Readiness Indicator file check\\\\n2025-12-06T03:35:15Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bdx7g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:23Z\\\"}}\" for pod \"openshift-multus\"/\"multus-kq4z8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:35:16Z is after 2025-08-24T17:21:41Z" Dec 06 03:35:16 crc kubenswrapper[4980]: I1206 03:35:16.399067 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-6kxfq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"951b6deb-bcea-4332-8ae8-c23ac7aaef36\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://423d30c3e65b04bdf6ff5ef53d8a7575a434c05e28489370a7370381ebc29de9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b8htf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://282df50dd9653c1cbe97fc2041d9e9ea91a992661682652aafc12e566d8f1b81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b8htf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:37Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-6kxfq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:35:16Z is after 2025-08-24T17:21:41Z" Dec 06 03:35:16 crc kubenswrapper[4980]: I1206 03:35:16.413847 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-mcprh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e6e88c5e-a255-449a-ae38-ac4d73a8e984\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g7xz9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g7xz9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:38Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-mcprh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:35:16Z is after 2025-08-24T17:21:41Z" Dec 06 03:35:16 crc kubenswrapper[4980]: I1206 03:35:16.425988 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:35:16Z is after 2025-08-24T17:21:41Z" Dec 06 03:35:16 crc kubenswrapper[4980]: I1206 03:35:16.437914 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4ef01e7d9809fd9cfc46e3a58b07159cdc6f5a3c210d1c9efc49bca27b30858\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:35:16Z is after 2025-08-24T17:21:41Z" Dec 06 03:35:16 crc kubenswrapper[4980]: I1206 03:35:16.483924 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:16 crc kubenswrapper[4980]: I1206 03:35:16.484008 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:16 crc kubenswrapper[4980]: I1206 03:35:16.484018 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:16 crc kubenswrapper[4980]: I1206 03:35:16.484031 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:16 crc kubenswrapper[4980]: I1206 03:35:16.484041 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:16Z","lastTransitionTime":"2025-12-06T03:35:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:16 crc kubenswrapper[4980]: I1206 03:35:16.586810 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:16 crc kubenswrapper[4980]: I1206 03:35:16.587023 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:16 crc kubenswrapper[4980]: I1206 03:35:16.587166 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:16 crc kubenswrapper[4980]: I1206 03:35:16.587247 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:16 crc kubenswrapper[4980]: I1206 03:35:16.587331 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:16Z","lastTransitionTime":"2025-12-06T03:35:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:16 crc kubenswrapper[4980]: I1206 03:35:16.690060 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:16 crc kubenswrapper[4980]: I1206 03:35:16.690105 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:16 crc kubenswrapper[4980]: I1206 03:35:16.690120 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:16 crc kubenswrapper[4980]: I1206 03:35:16.690136 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:16 crc kubenswrapper[4980]: I1206 03:35:16.690147 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:16Z","lastTransitionTime":"2025-12-06T03:35:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:16 crc kubenswrapper[4980]: I1206 03:35:16.792728 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:16 crc kubenswrapper[4980]: I1206 03:35:16.792816 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:16 crc kubenswrapper[4980]: I1206 03:35:16.792840 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:16 crc kubenswrapper[4980]: I1206 03:35:16.792878 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:16 crc kubenswrapper[4980]: I1206 03:35:16.792921 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:16Z","lastTransitionTime":"2025-12-06T03:35:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:16 crc kubenswrapper[4980]: I1206 03:35:16.895804 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:16 crc kubenswrapper[4980]: I1206 03:35:16.895885 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:16 crc kubenswrapper[4980]: I1206 03:35:16.895908 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:16 crc kubenswrapper[4980]: I1206 03:35:16.895936 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:16 crc kubenswrapper[4980]: I1206 03:35:16.895954 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:16Z","lastTransitionTime":"2025-12-06T03:35:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:17 crc kubenswrapper[4980]: I1206 03:35:16.999237 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:17 crc kubenswrapper[4980]: I1206 03:35:16.999336 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:17 crc kubenswrapper[4980]: I1206 03:35:16.999362 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:17 crc kubenswrapper[4980]: I1206 03:35:16.999398 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:17 crc kubenswrapper[4980]: I1206 03:35:16.999432 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:16Z","lastTransitionTime":"2025-12-06T03:35:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:17 crc kubenswrapper[4980]: I1206 03:35:17.102297 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:17 crc kubenswrapper[4980]: I1206 03:35:17.102335 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:17 crc kubenswrapper[4980]: I1206 03:35:17.102345 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:17 crc kubenswrapper[4980]: I1206 03:35:17.102360 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:17 crc kubenswrapper[4980]: I1206 03:35:17.102370 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:17Z","lastTransitionTime":"2025-12-06T03:35:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:17 crc kubenswrapper[4980]: I1206 03:35:17.119738 4980 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-kq4z8_f96c1f6c-55e8-4d86-bff3-5381581b7022/kube-multus/0.log" Dec 06 03:35:17 crc kubenswrapper[4980]: I1206 03:35:17.119808 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-kq4z8" event={"ID":"f96c1f6c-55e8-4d86-bff3-5381581b7022","Type":"ContainerStarted","Data":"ea7f0bc23247b5effe7f0250582c8694bbf0e9a2d185a844773665a0a48f350e"} Dec 06 03:35:17 crc kubenswrapper[4980]: I1206 03:35:17.137272 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4ef01e7d9809fd9cfc46e3a58b07159cdc6f5a3c210d1c9efc49bca27b30858\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:35:17Z is after 2025-08-24T17:21:41Z" Dec 06 03:35:17 crc kubenswrapper[4980]: I1206 03:35:17.151976 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-kq4z8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f96c1f6c-55e8-4d86-bff3-5381581b7022\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:35:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:35:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ea7f0bc23247b5effe7f0250582c8694bbf0e9a2d185a844773665a0a48f350e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://799498681afa7653408bc23b1e82d75c16e3c809d2ee7f322780e20c537eb710\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-06T03:35:15Z\\\",\\\"message\\\":\\\"2025-12-06T03:34:30+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_68ae7793-ff25-4e9c-8158-bea74df4ae16\\\\n2025-12-06T03:34:30+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_68ae7793-ff25-4e9c-8158-bea74df4ae16 to /host/opt/cni/bin/\\\\n2025-12-06T03:34:30Z [verbose] multus-daemon started\\\\n2025-12-06T03:34:30Z [verbose] Readiness Indicator file check\\\\n2025-12-06T03:35:15Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:24Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:35:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bdx7g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:23Z\\\"}}\" for pod \"openshift-multus\"/\"multus-kq4z8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:35:17Z is after 2025-08-24T17:21:41Z" Dec 06 03:35:17 crc kubenswrapper[4980]: I1206 03:35:17.161719 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-6kxfq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"951b6deb-bcea-4332-8ae8-c23ac7aaef36\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://423d30c3e65b04bdf6ff5ef53d8a7575a434c05e28489370a7370381ebc29de9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b8htf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://282df50dd9653c1cbe97fc2041d9e9ea91a992661682652aafc12e566d8f1b81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b8htf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:37Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-6kxfq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:35:17Z is after 2025-08-24T17:21:41Z" Dec 06 03:35:17 crc kubenswrapper[4980]: I1206 03:35:17.171308 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-mcprh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e6e88c5e-a255-449a-ae38-ac4d73a8e984\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g7xz9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g7xz9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:38Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-mcprh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:35:17Z is after 2025-08-24T17:21:41Z" Dec 06 03:35:17 crc kubenswrapper[4980]: I1206 03:35:17.181590 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:35:17Z is after 2025-08-24T17:21:41Z" Dec 06 03:35:17 crc kubenswrapper[4980]: I1206 03:35:17.183621 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 03:35:17 crc kubenswrapper[4980]: E1206 03:35:17.183765 4980 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 03:35:17 crc kubenswrapper[4980]: I1206 03:35:17.198953 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7f958150-549d-4c80-8a7e-23c39e4f9500\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c28b6895b14e73f4f3eb7a8ce93076c05092ba9b58ea0e3e879c65bcfa227a21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://76cff8a2d652df38b04eefc053832baf4cbb5db778ff5e525b18c2825141b118\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://206f6f19fe3197f5d6798cc86c5473bf2decb02a2860ee7ffb5efb5f60952e8c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bafaa9670b05671fb57b658c2960bf3ee32c84e52f86a3a3bbab01d5a38ff0c4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://773a0bc2aab73ffc2717e987543defc9b786cc22300aa54b4ba4f3454504e331\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-06T03:34:22Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1206 03:34:15.305330 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1206 03:34:15.307249 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1592846546/tls.crt::/tmp/serving-cert-1592846546/tls.key\\\\\\\"\\\\nI1206 03:34:21.814037 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1206 03:34:21.816756 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1206 03:34:21.816780 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1206 03:34:21.816806 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1206 03:34:21.816813 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1206 03:34:21.831301 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1206 03:34:21.831311 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1206 03:34:21.831324 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 03:34:21.831330 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 03:34:21.831335 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1206 03:34:21.831339 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1206 03:34:21.831342 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1206 03:34:21.831346 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1206 03:34:21.911269 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:04Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d40e0e7446f8416bf9da94fc53c01be937f9a51a11eb101157d2ddd4597720d2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:04Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2cba89f2f1bed96d59f183628d900d9b72c91eb472e273c8bed0babcab485910\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2cba89f2f1bed96d59f183628d900d9b72c91eb472e273c8bed0babcab485910\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:01Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:35:17Z is after 2025-08-24T17:21:41Z" Dec 06 03:35:17 crc kubenswrapper[4980]: I1206 03:35:17.205311 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:17 crc kubenswrapper[4980]: I1206 03:35:17.205352 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:17 crc kubenswrapper[4980]: I1206 03:35:17.205367 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:17 crc kubenswrapper[4980]: I1206 03:35:17.205386 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:17 crc kubenswrapper[4980]: I1206 03:35:17.205401 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:17Z","lastTransitionTime":"2025-12-06T03:35:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:17 crc kubenswrapper[4980]: I1206 03:35:17.249785 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6ed68f66-1cf6-465a-8f66-76d5af6a44d8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://725e84f6bdfb722a8ce69158d03ea4c66a9f087642d55cb9651e003c81be0c8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://71f7b492ea07cdf7dc65fc48184ec1e64914eb1811bbfdddc0fb14c8ad6cb296\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b543aa4ee03d5a14909b67d876c5675c02e1df2a89a3c9ac8d3fb587fea53d8f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://32af3427e872a70cb3c2357116f04cabc8d3d9b69e7306b2e0de2ae22f9469d4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:01Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:35:17Z is after 2025-08-24T17:21:41Z" Dec 06 03:35:17 crc kubenswrapper[4980]: I1206 03:35:17.262983 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:22Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:35:17Z is after 2025-08-24T17:21:41Z" Dec 06 03:35:17 crc kubenswrapper[4980]: I1206 03:35:17.274297 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://334ea6a273fd7d268906c9f4a6dbf6f4df4fcdbf38f3d7d248cb7d6cf8537794\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ec44dcd5e4bf0a0466bb586ebd61259364dbbfe22499f4cbff854e3bd153082\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:35:17Z is after 2025-08-24T17:21:41Z" Dec 06 03:35:17 crc kubenswrapper[4980]: I1206 03:35:17.283875 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:22Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:35:17Z is after 2025-08-24T17:21:41Z" Dec 06 03:35:17 crc kubenswrapper[4980]: I1206 03:35:17.293284 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-sv5sz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"873977aa-4eb0-4526-bc0b-757ad2dea2a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01f0cc855c822a71ea8ef6a7ce6e25747ff9956691735973f935c4a28672a7a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lg7tx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:23Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-sv5sz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:35:17Z is after 2025-08-24T17:21:41Z" Dec 06 03:35:17 crc kubenswrapper[4980]: I1206 03:35:17.307971 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:17 crc kubenswrapper[4980]: I1206 03:35:17.308002 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:17 crc kubenswrapper[4980]: I1206 03:35:17.308013 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:17 crc kubenswrapper[4980]: I1206 03:35:17.308027 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:17 crc kubenswrapper[4980]: I1206 03:35:17.308037 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:17Z","lastTransitionTime":"2025-12-06T03:35:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:17 crc kubenswrapper[4980]: I1206 03:35:17.315757 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5w4fr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5af0e768-3c9c-4401-ab56-57516cd3170e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:23Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:23Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://768b8e1fe1a00a8467d744da5ad3f61006b0a3612c96f8e4d0ee6fcb93195f9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfzrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6342903b318317ee8d84c2f5d467cb25c811a789be0d6ab0e0cba37b5611e16d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfzrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d620685574db7594132ae24a91ec0a93dce20a9a289540deebb39650b8a1816\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfzrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c67ec8f9c1764fcdf215d52822c2d4f0437dc7f056f05d708767121f6c1f3515\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfzrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f35beae5339eeaf26494ed678e3c1bd02349eae2a4b4615325b17923f3b653b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfzrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a940ec6ee523313728f7e8ae934b7e0b0cff82771e3e64723309db1fae013c3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfzrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fae428d1a6eaeb2e6ddd4a5b1a1170e230635739ea0b39743e565e47fb7ee6e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fae428d1a6eaeb2e6ddd4a5b1a1170e230635739ea0b39743e565e47fb7ee6e4\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-06T03:34:53Z\\\",\\\"message\\\":\\\"/pkg/client/informers/externalversions/factory.go:141\\\\nI1206 03:34:53.908007 6569 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1206 03:34:53.908049 6569 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1206 03:34:53.908094 6569 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1206 03:34:53.908109 6569 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1206 03:34:53.908160 6569 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1206 03:34:53.908208 6569 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1206 03:34:53.908256 6569 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1206 03:34:53.908286 6569 factory.go:656] Stopping watch factory\\\\nI1206 03:34:53.908308 6569 ovnkube.go:599] Stopped ovnkube\\\\nI1206 03:34:53.908349 6569 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1206 03:34:53.908364 6569 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1206 03:34:53.908372 6569 handler.go:208] Removed *v1.Node event handler 2\\\\nI1206 03:34:53.908385 6569 handler.go:208] Removed *v1.Node event handler 7\\\\nI1206 03:34:53.908397 6569 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1206 03:34:53.908415 6569 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1206 03:34:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:52Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-5w4fr_openshift-ovn-kubernetes(5af0e768-3c9c-4401-ab56-57516cd3170e)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfzrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe261f4322b85702181d7473ab2a9a4d2d4b869d586c3091f811b12a81a63aae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfzrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d876df04b5ab0ba07a5cbcafbe4648403c5014c59eb7d40adee2657c3a255473\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d876df04b5ab0ba07a5cbcafbe4648403c5014c59eb7d40adee2657c3a255473\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfzrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:23Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-5w4fr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:35:17Z is after 2025-08-24T17:21:41Z" Dec 06 03:35:17 crc kubenswrapper[4980]: I1206 03:35:17.326920 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"64278250-0b19-4591-b06e-6c96652ee92a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5beac3fd93b448506c0a110d70ef5faf42efa18b16abe6b3d4eb7b9c4233ad93\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ae66052702fcfd94f0123f4fbbe9efed37eb04609269a2bc951f95669de16601\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://173903f3554d7ac9158493498a016d618d92f46da1500408834b3c24249cf117\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1f596bfd5321ca225f4be03d4cf229c31a52def9731c0e177ab3a0bc010380e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1f596bfd5321ca225f4be03d4cf229c31a52def9731c0e177ab3a0bc010380e1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:02Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:01Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:35:17Z is after 2025-08-24T17:21:41Z" Dec 06 03:35:17 crc kubenswrapper[4980]: I1206 03:35:17.337324 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://502c727d6baa31f400bee14c34b19abbe38af873f28f6a82a43547431755d981\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:35:17Z is after 2025-08-24T17:21:41Z" Dec 06 03:35:17 crc kubenswrapper[4980]: I1206 03:35:17.349883 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-dx44q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5bac22a-512c-49d0-90db-47de6aaaefd0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://97469dea2cd5393da0acf784ccc1969c8a4d07da7d51b14cdc955013ee9710ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f7mdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3b5dab476c7f227a6e4a2eae8469432667fb9fad345d9419ba59feab9fd385e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c3b5dab476c7f227a6e4a2eae8469432667fb9fad345d9419ba59feab9fd385e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f7mdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e921499231cbdf26b8918c98f9ad8c322fd7848865caeef048470de3e2c0efba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e921499231cbdf26b8918c98f9ad8c322fd7848865caeef048470de3e2c0efba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f7mdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7cbfe7d8cdb89916c238296592fab45b9e92e2c585baa6d67230decd1c84c5bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7cbfe7d8cdb89916c238296592fab45b9e92e2c585baa6d67230decd1c84c5bb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f7mdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f45cbf23b6b0d208cefc4b493f901bab9d71ccf2f1456507d54fb1f30c4fd48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7f45cbf23b6b0d208cefc4b493f901bab9d71ccf2f1456507d54fb1f30c4fd48\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f7mdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://344ce4b145ff021fcaf163803d8ad8cc6e9565bf7f0992e2653c43db663d8411\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://344ce4b145ff021fcaf163803d8ad8cc6e9565bf7f0992e2653c43db663d8411\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f7mdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://619e56a9937dd6376ab471ba16da30e5d77f443f5bd33610bea5ec85cd0446ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://619e56a9937dd6376ab471ba16da30e5d77f443f5bd33610bea5ec85cd0446ef\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f7mdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:23Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-dx44q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:35:17Z is after 2025-08-24T17:21:41Z" Dec 06 03:35:17 crc kubenswrapper[4980]: I1206 03:35:17.359429 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-r5zfb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"320f44d1-a671-4a91-b328-a8b0fdd8f23a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://92393ecaaef8cfcfd01837671f8cc89803dd570b637b4a373b40b7f487788826\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zhm5t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52e8db644463794041978ced2c0591c13e0ada8e5f1a78a94a0a665218a37b5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zhm5t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:23Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-r5zfb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:35:17Z is after 2025-08-24T17:21:41Z" Dec 06 03:35:17 crc kubenswrapper[4980]: I1206 03:35:17.367777 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-cvzg8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f1cde706-268d-4d26-818f-2f42bfb37b5f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://86f9a77acae360740be827eb0d1e1516e8a49cb9b197b6f5f9468ac70a101158\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rk86l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:29Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-cvzg8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:35:17Z is after 2025-08-24T17:21:41Z" Dec 06 03:35:17 crc kubenswrapper[4980]: I1206 03:35:17.391033 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cab608b3-a02f-475b-9aae-9b8f5d878de9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f7fcc44b42bcbb55509048e3a566bb557333c4a83e33d7b0fba8eb514317909\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c22e7dd8886c2dfbb742900ed86ec7b0c2edf9a1363e74f6666961d928f06c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://473e792b2d0a9644854ee0581c6120fc9c989bc342899ef72b1d056080fd4694\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ee8e936dfc39361d6f99460910591ca8c901a20f0d274bbb6aaebfa8a00ee06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://77bc12768b018625ba11a6bfd66688f926c4b7f3bc1d0e41c6551a18b485ecf3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://27cf07cd96893900da4afecd0d56fd9269bc60c1d23246b4685fcd8756357c19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://27cf07cd96893900da4afecd0d56fd9269bc60c1d23246b4685fcd8756357c19\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c49feb24566a8284378670fa0390710d38c237acc94ffb654f90f3fa4626e4c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c49feb24566a8284378670fa0390710d38c237acc94ffb654f90f3fa4626e4c8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:03Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://bb74d1c3d752a379fa394c450217badb45511fbcafad62f7b618131cab94b29e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bb74d1c3d752a379fa394c450217badb45511fbcafad62f7b618131cab94b29e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:01Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:35:17Z is after 2025-08-24T17:21:41Z" Dec 06 03:35:17 crc kubenswrapper[4980]: I1206 03:35:17.411364 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:17 crc kubenswrapper[4980]: I1206 03:35:17.411412 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:17 crc kubenswrapper[4980]: I1206 03:35:17.411424 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:17 crc kubenswrapper[4980]: I1206 03:35:17.411443 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:17 crc kubenswrapper[4980]: I1206 03:35:17.411457 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:17Z","lastTransitionTime":"2025-12-06T03:35:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:17 crc kubenswrapper[4980]: I1206 03:35:17.514770 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:17 crc kubenswrapper[4980]: I1206 03:35:17.514815 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:17 crc kubenswrapper[4980]: I1206 03:35:17.514824 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:17 crc kubenswrapper[4980]: I1206 03:35:17.514838 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:17 crc kubenswrapper[4980]: I1206 03:35:17.514848 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:17Z","lastTransitionTime":"2025-12-06T03:35:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:17 crc kubenswrapper[4980]: I1206 03:35:17.618209 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:17 crc kubenswrapper[4980]: I1206 03:35:17.618267 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:17 crc kubenswrapper[4980]: I1206 03:35:17.618286 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:17 crc kubenswrapper[4980]: I1206 03:35:17.618310 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:17 crc kubenswrapper[4980]: I1206 03:35:17.618332 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:17Z","lastTransitionTime":"2025-12-06T03:35:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:17 crc kubenswrapper[4980]: I1206 03:35:17.721369 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:17 crc kubenswrapper[4980]: I1206 03:35:17.721437 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:17 crc kubenswrapper[4980]: I1206 03:35:17.721454 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:17 crc kubenswrapper[4980]: I1206 03:35:17.721480 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:17 crc kubenswrapper[4980]: I1206 03:35:17.721500 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:17Z","lastTransitionTime":"2025-12-06T03:35:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:17 crc kubenswrapper[4980]: I1206 03:35:17.824179 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:17 crc kubenswrapper[4980]: I1206 03:35:17.824247 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:17 crc kubenswrapper[4980]: I1206 03:35:17.824265 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:17 crc kubenswrapper[4980]: I1206 03:35:17.824296 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:17 crc kubenswrapper[4980]: I1206 03:35:17.824319 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:17Z","lastTransitionTime":"2025-12-06T03:35:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:17 crc kubenswrapper[4980]: I1206 03:35:17.928141 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:17 crc kubenswrapper[4980]: I1206 03:35:17.928233 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:17 crc kubenswrapper[4980]: I1206 03:35:17.928260 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:17 crc kubenswrapper[4980]: I1206 03:35:17.928292 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:17 crc kubenswrapper[4980]: I1206 03:35:17.928317 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:17Z","lastTransitionTime":"2025-12-06T03:35:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:18 crc kubenswrapper[4980]: I1206 03:35:18.032298 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:18 crc kubenswrapper[4980]: I1206 03:35:18.032364 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:18 crc kubenswrapper[4980]: I1206 03:35:18.032382 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:18 crc kubenswrapper[4980]: I1206 03:35:18.032408 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:18 crc kubenswrapper[4980]: I1206 03:35:18.032429 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:18Z","lastTransitionTime":"2025-12-06T03:35:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:18 crc kubenswrapper[4980]: I1206 03:35:18.134988 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:18 crc kubenswrapper[4980]: I1206 03:35:18.135035 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:18 crc kubenswrapper[4980]: I1206 03:35:18.135047 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:18 crc kubenswrapper[4980]: I1206 03:35:18.135062 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:18 crc kubenswrapper[4980]: I1206 03:35:18.135075 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:18Z","lastTransitionTime":"2025-12-06T03:35:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:18 crc kubenswrapper[4980]: I1206 03:35:18.184204 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 03:35:18 crc kubenswrapper[4980]: I1206 03:35:18.184260 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-mcprh" Dec 06 03:35:18 crc kubenswrapper[4980]: I1206 03:35:18.184334 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 03:35:18 crc kubenswrapper[4980]: E1206 03:35:18.184389 4980 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 03:35:18 crc kubenswrapper[4980]: E1206 03:35:18.184566 4980 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-mcprh" podUID="e6e88c5e-a255-449a-ae38-ac4d73a8e984" Dec 06 03:35:18 crc kubenswrapper[4980]: E1206 03:35:18.184677 4980 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 03:35:18 crc kubenswrapper[4980]: I1206 03:35:18.239278 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:18 crc kubenswrapper[4980]: I1206 03:35:18.239338 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:18 crc kubenswrapper[4980]: I1206 03:35:18.239353 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:18 crc kubenswrapper[4980]: I1206 03:35:18.239382 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:18 crc kubenswrapper[4980]: I1206 03:35:18.239397 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:18Z","lastTransitionTime":"2025-12-06T03:35:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:18 crc kubenswrapper[4980]: I1206 03:35:18.343371 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:18 crc kubenswrapper[4980]: I1206 03:35:18.343494 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:18 crc kubenswrapper[4980]: I1206 03:35:18.343562 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:18 crc kubenswrapper[4980]: I1206 03:35:18.343596 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:18 crc kubenswrapper[4980]: I1206 03:35:18.343623 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:18Z","lastTransitionTime":"2025-12-06T03:35:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:18 crc kubenswrapper[4980]: I1206 03:35:18.446975 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:18 crc kubenswrapper[4980]: I1206 03:35:18.447050 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:18 crc kubenswrapper[4980]: I1206 03:35:18.447087 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:18 crc kubenswrapper[4980]: I1206 03:35:18.447115 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:18 crc kubenswrapper[4980]: I1206 03:35:18.447137 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:18Z","lastTransitionTime":"2025-12-06T03:35:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:18 crc kubenswrapper[4980]: I1206 03:35:18.550535 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:18 crc kubenswrapper[4980]: I1206 03:35:18.550579 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:18 crc kubenswrapper[4980]: I1206 03:35:18.550591 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:18 crc kubenswrapper[4980]: I1206 03:35:18.550608 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:18 crc kubenswrapper[4980]: I1206 03:35:18.550618 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:18Z","lastTransitionTime":"2025-12-06T03:35:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:18 crc kubenswrapper[4980]: I1206 03:35:18.653539 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:18 crc kubenswrapper[4980]: I1206 03:35:18.653606 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:18 crc kubenswrapper[4980]: I1206 03:35:18.653624 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:18 crc kubenswrapper[4980]: I1206 03:35:18.653651 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:18 crc kubenswrapper[4980]: I1206 03:35:18.653670 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:18Z","lastTransitionTime":"2025-12-06T03:35:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:18 crc kubenswrapper[4980]: I1206 03:35:18.757470 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:18 crc kubenswrapper[4980]: I1206 03:35:18.757552 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:18 crc kubenswrapper[4980]: I1206 03:35:18.757568 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:18 crc kubenswrapper[4980]: I1206 03:35:18.757589 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:18 crc kubenswrapper[4980]: I1206 03:35:18.757605 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:18Z","lastTransitionTime":"2025-12-06T03:35:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:18 crc kubenswrapper[4980]: I1206 03:35:18.861561 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:18 crc kubenswrapper[4980]: I1206 03:35:18.861649 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:18 crc kubenswrapper[4980]: I1206 03:35:18.861670 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:18 crc kubenswrapper[4980]: I1206 03:35:18.861696 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:18 crc kubenswrapper[4980]: I1206 03:35:18.861713 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:18Z","lastTransitionTime":"2025-12-06T03:35:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:18 crc kubenswrapper[4980]: I1206 03:35:18.965094 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:18 crc kubenswrapper[4980]: I1206 03:35:18.965159 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:18 crc kubenswrapper[4980]: I1206 03:35:18.965178 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:18 crc kubenswrapper[4980]: I1206 03:35:18.965207 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:18 crc kubenswrapper[4980]: I1206 03:35:18.965226 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:18Z","lastTransitionTime":"2025-12-06T03:35:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:19 crc kubenswrapper[4980]: I1206 03:35:19.069120 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:19 crc kubenswrapper[4980]: I1206 03:35:19.069182 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:19 crc kubenswrapper[4980]: I1206 03:35:19.069198 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:19 crc kubenswrapper[4980]: I1206 03:35:19.069228 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:19 crc kubenswrapper[4980]: I1206 03:35:19.069247 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:19Z","lastTransitionTime":"2025-12-06T03:35:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:19 crc kubenswrapper[4980]: I1206 03:35:19.172407 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:19 crc kubenswrapper[4980]: I1206 03:35:19.172475 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:19 crc kubenswrapper[4980]: I1206 03:35:19.172494 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:19 crc kubenswrapper[4980]: I1206 03:35:19.172548 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:19 crc kubenswrapper[4980]: I1206 03:35:19.172567 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:19Z","lastTransitionTime":"2025-12-06T03:35:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:19 crc kubenswrapper[4980]: I1206 03:35:19.183942 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 03:35:19 crc kubenswrapper[4980]: E1206 03:35:19.184081 4980 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 03:35:19 crc kubenswrapper[4980]: I1206 03:35:19.276185 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:19 crc kubenswrapper[4980]: I1206 03:35:19.276250 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:19 crc kubenswrapper[4980]: I1206 03:35:19.276261 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:19 crc kubenswrapper[4980]: I1206 03:35:19.276279 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:19 crc kubenswrapper[4980]: I1206 03:35:19.276294 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:19Z","lastTransitionTime":"2025-12-06T03:35:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:19 crc kubenswrapper[4980]: I1206 03:35:19.378713 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:19 crc kubenswrapper[4980]: I1206 03:35:19.378781 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:19 crc kubenswrapper[4980]: I1206 03:35:19.378807 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:19 crc kubenswrapper[4980]: I1206 03:35:19.378838 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:19 crc kubenswrapper[4980]: I1206 03:35:19.378860 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:19Z","lastTransitionTime":"2025-12-06T03:35:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:19 crc kubenswrapper[4980]: I1206 03:35:19.482177 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:19 crc kubenswrapper[4980]: I1206 03:35:19.482246 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:19 crc kubenswrapper[4980]: I1206 03:35:19.482262 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:19 crc kubenswrapper[4980]: I1206 03:35:19.482281 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:19 crc kubenswrapper[4980]: I1206 03:35:19.482294 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:19Z","lastTransitionTime":"2025-12-06T03:35:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:19 crc kubenswrapper[4980]: I1206 03:35:19.585158 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:19 crc kubenswrapper[4980]: I1206 03:35:19.585209 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:19 crc kubenswrapper[4980]: I1206 03:35:19.585221 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:19 crc kubenswrapper[4980]: I1206 03:35:19.585240 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:19 crc kubenswrapper[4980]: I1206 03:35:19.585252 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:19Z","lastTransitionTime":"2025-12-06T03:35:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:19 crc kubenswrapper[4980]: I1206 03:35:19.688253 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:19 crc kubenswrapper[4980]: I1206 03:35:19.688439 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:19 crc kubenswrapper[4980]: I1206 03:35:19.688464 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:19 crc kubenswrapper[4980]: I1206 03:35:19.688561 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:19 crc kubenswrapper[4980]: I1206 03:35:19.688590 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:19Z","lastTransitionTime":"2025-12-06T03:35:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:19 crc kubenswrapper[4980]: I1206 03:35:19.791887 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:19 crc kubenswrapper[4980]: I1206 03:35:19.791970 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:19 crc kubenswrapper[4980]: I1206 03:35:19.791981 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:19 crc kubenswrapper[4980]: I1206 03:35:19.792000 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:19 crc kubenswrapper[4980]: I1206 03:35:19.792012 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:19Z","lastTransitionTime":"2025-12-06T03:35:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:19 crc kubenswrapper[4980]: I1206 03:35:19.857417 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:19 crc kubenswrapper[4980]: I1206 03:35:19.857492 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:19 crc kubenswrapper[4980]: I1206 03:35:19.857604 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:19 crc kubenswrapper[4980]: I1206 03:35:19.857641 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:19 crc kubenswrapper[4980]: I1206 03:35:19.857663 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:19Z","lastTransitionTime":"2025-12-06T03:35:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:19 crc kubenswrapper[4980]: E1206 03:35:19.884657 4980 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T03:35:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T03:35:19Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T03:35:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T03:35:19Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T03:35:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T03:35:19Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T03:35:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T03:35:19Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"9b4ed4f8-ed8d-447b-9e1f-13defbdcf0f7\\\",\\\"systemUUID\\\":\\\"a5b493aa-69de-4688-8ff6-1e5f49b7f014\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:35:19Z is after 2025-08-24T17:21:41Z" Dec 06 03:35:19 crc kubenswrapper[4980]: I1206 03:35:19.889304 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:19 crc kubenswrapper[4980]: I1206 03:35:19.889358 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:19 crc kubenswrapper[4980]: I1206 03:35:19.889371 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:19 crc kubenswrapper[4980]: I1206 03:35:19.889389 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:19 crc kubenswrapper[4980]: I1206 03:35:19.889402 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:19Z","lastTransitionTime":"2025-12-06T03:35:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:19 crc kubenswrapper[4980]: E1206 03:35:19.905441 4980 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T03:35:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T03:35:19Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T03:35:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T03:35:19Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T03:35:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T03:35:19Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T03:35:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T03:35:19Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"9b4ed4f8-ed8d-447b-9e1f-13defbdcf0f7\\\",\\\"systemUUID\\\":\\\"a5b493aa-69de-4688-8ff6-1e5f49b7f014\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:35:19Z is after 2025-08-24T17:21:41Z" Dec 06 03:35:19 crc kubenswrapper[4980]: I1206 03:35:19.908822 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:19 crc kubenswrapper[4980]: I1206 03:35:19.908859 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:19 crc kubenswrapper[4980]: I1206 03:35:19.908871 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:19 crc kubenswrapper[4980]: I1206 03:35:19.908887 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:19 crc kubenswrapper[4980]: I1206 03:35:19.908900 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:19Z","lastTransitionTime":"2025-12-06T03:35:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:19 crc kubenswrapper[4980]: E1206 03:35:19.931034 4980 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T03:35:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T03:35:19Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T03:35:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T03:35:19Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T03:35:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T03:35:19Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T03:35:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T03:35:19Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"9b4ed4f8-ed8d-447b-9e1f-13defbdcf0f7\\\",\\\"systemUUID\\\":\\\"a5b493aa-69de-4688-8ff6-1e5f49b7f014\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:35:19Z is after 2025-08-24T17:21:41Z" Dec 06 03:35:19 crc kubenswrapper[4980]: I1206 03:35:19.934832 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:19 crc kubenswrapper[4980]: I1206 03:35:19.934867 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:19 crc kubenswrapper[4980]: I1206 03:35:19.934894 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:19 crc kubenswrapper[4980]: I1206 03:35:19.934918 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:19 crc kubenswrapper[4980]: I1206 03:35:19.934933 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:19Z","lastTransitionTime":"2025-12-06T03:35:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:19 crc kubenswrapper[4980]: E1206 03:35:19.948843 4980 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T03:35:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T03:35:19Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T03:35:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T03:35:19Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T03:35:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T03:35:19Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T03:35:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T03:35:19Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"9b4ed4f8-ed8d-447b-9e1f-13defbdcf0f7\\\",\\\"systemUUID\\\":\\\"a5b493aa-69de-4688-8ff6-1e5f49b7f014\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:35:19Z is after 2025-08-24T17:21:41Z" Dec 06 03:35:19 crc kubenswrapper[4980]: I1206 03:35:19.952686 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:19 crc kubenswrapper[4980]: I1206 03:35:19.952733 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:19 crc kubenswrapper[4980]: I1206 03:35:19.952941 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:19 crc kubenswrapper[4980]: I1206 03:35:19.952962 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:19 crc kubenswrapper[4980]: I1206 03:35:19.952974 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:19Z","lastTransitionTime":"2025-12-06T03:35:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:19 crc kubenswrapper[4980]: E1206 03:35:19.964968 4980 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T03:35:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T03:35:19Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T03:35:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T03:35:19Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T03:35:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T03:35:19Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T03:35:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T03:35:19Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"9b4ed4f8-ed8d-447b-9e1f-13defbdcf0f7\\\",\\\"systemUUID\\\":\\\"a5b493aa-69de-4688-8ff6-1e5f49b7f014\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:35:19Z is after 2025-08-24T17:21:41Z" Dec 06 03:35:19 crc kubenswrapper[4980]: E1206 03:35:19.965097 4980 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 06 03:35:19 crc kubenswrapper[4980]: I1206 03:35:19.966310 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:19 crc kubenswrapper[4980]: I1206 03:35:19.966341 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:19 crc kubenswrapper[4980]: I1206 03:35:19.966354 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:19 crc kubenswrapper[4980]: I1206 03:35:19.966372 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:19 crc kubenswrapper[4980]: I1206 03:35:19.966383 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:19Z","lastTransitionTime":"2025-12-06T03:35:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:20 crc kubenswrapper[4980]: I1206 03:35:20.069029 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:20 crc kubenswrapper[4980]: I1206 03:35:20.069060 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:20 crc kubenswrapper[4980]: I1206 03:35:20.069070 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:20 crc kubenswrapper[4980]: I1206 03:35:20.069088 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:20 crc kubenswrapper[4980]: I1206 03:35:20.069099 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:20Z","lastTransitionTime":"2025-12-06T03:35:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:20 crc kubenswrapper[4980]: I1206 03:35:20.172167 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:20 crc kubenswrapper[4980]: I1206 03:35:20.172218 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:20 crc kubenswrapper[4980]: I1206 03:35:20.172230 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:20 crc kubenswrapper[4980]: I1206 03:35:20.172251 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:20 crc kubenswrapper[4980]: I1206 03:35:20.172264 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:20Z","lastTransitionTime":"2025-12-06T03:35:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:20 crc kubenswrapper[4980]: I1206 03:35:20.183858 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-mcprh" Dec 06 03:35:20 crc kubenswrapper[4980]: I1206 03:35:20.184080 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 03:35:20 crc kubenswrapper[4980]: I1206 03:35:20.184114 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 03:35:20 crc kubenswrapper[4980]: E1206 03:35:20.184201 4980 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 03:35:20 crc kubenswrapper[4980]: I1206 03:35:20.184420 4980 scope.go:117] "RemoveContainer" containerID="fae428d1a6eaeb2e6ddd4a5b1a1170e230635739ea0b39743e565e47fb7ee6e4" Dec 06 03:35:20 crc kubenswrapper[4980]: E1206 03:35:20.184491 4980 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 03:35:20 crc kubenswrapper[4980]: E1206 03:35:20.184673 4980 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-mcprh" podUID="e6e88c5e-a255-449a-ae38-ac4d73a8e984" Dec 06 03:35:20 crc kubenswrapper[4980]: I1206 03:35:20.275613 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:20 crc kubenswrapper[4980]: I1206 03:35:20.275657 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:20 crc kubenswrapper[4980]: I1206 03:35:20.275669 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:20 crc kubenswrapper[4980]: I1206 03:35:20.275689 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:20 crc kubenswrapper[4980]: I1206 03:35:20.275705 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:20Z","lastTransitionTime":"2025-12-06T03:35:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:20 crc kubenswrapper[4980]: I1206 03:35:20.378442 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:20 crc kubenswrapper[4980]: I1206 03:35:20.378479 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:20 crc kubenswrapper[4980]: I1206 03:35:20.378488 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:20 crc kubenswrapper[4980]: I1206 03:35:20.378503 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:20 crc kubenswrapper[4980]: I1206 03:35:20.378525 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:20Z","lastTransitionTime":"2025-12-06T03:35:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:20 crc kubenswrapper[4980]: I1206 03:35:20.481549 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:20 crc kubenswrapper[4980]: I1206 03:35:20.481615 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:20 crc kubenswrapper[4980]: I1206 03:35:20.481626 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:20 crc kubenswrapper[4980]: I1206 03:35:20.481640 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:20 crc kubenswrapper[4980]: I1206 03:35:20.481650 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:20Z","lastTransitionTime":"2025-12-06T03:35:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:20 crc kubenswrapper[4980]: I1206 03:35:20.584207 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:20 crc kubenswrapper[4980]: I1206 03:35:20.584239 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:20 crc kubenswrapper[4980]: I1206 03:35:20.584246 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:20 crc kubenswrapper[4980]: I1206 03:35:20.584259 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:20 crc kubenswrapper[4980]: I1206 03:35:20.584268 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:20Z","lastTransitionTime":"2025-12-06T03:35:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:20 crc kubenswrapper[4980]: I1206 03:35:20.686555 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:20 crc kubenswrapper[4980]: I1206 03:35:20.686596 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:20 crc kubenswrapper[4980]: I1206 03:35:20.686606 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:20 crc kubenswrapper[4980]: I1206 03:35:20.686622 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:20 crc kubenswrapper[4980]: I1206 03:35:20.686632 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:20Z","lastTransitionTime":"2025-12-06T03:35:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:20 crc kubenswrapper[4980]: I1206 03:35:20.789914 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:20 crc kubenswrapper[4980]: I1206 03:35:20.789962 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:20 crc kubenswrapper[4980]: I1206 03:35:20.789971 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:20 crc kubenswrapper[4980]: I1206 03:35:20.789986 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:20 crc kubenswrapper[4980]: I1206 03:35:20.789996 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:20Z","lastTransitionTime":"2025-12-06T03:35:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:20 crc kubenswrapper[4980]: I1206 03:35:20.892366 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:20 crc kubenswrapper[4980]: I1206 03:35:20.892397 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:20 crc kubenswrapper[4980]: I1206 03:35:20.892409 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:20 crc kubenswrapper[4980]: I1206 03:35:20.892425 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:20 crc kubenswrapper[4980]: I1206 03:35:20.892436 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:20Z","lastTransitionTime":"2025-12-06T03:35:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:20 crc kubenswrapper[4980]: I1206 03:35:20.995613 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:20 crc kubenswrapper[4980]: I1206 03:35:20.996062 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:20 crc kubenswrapper[4980]: I1206 03:35:20.996146 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:20 crc kubenswrapper[4980]: I1206 03:35:20.996212 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:20 crc kubenswrapper[4980]: I1206 03:35:20.996278 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:20Z","lastTransitionTime":"2025-12-06T03:35:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:21 crc kubenswrapper[4980]: I1206 03:35:21.098338 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:21 crc kubenswrapper[4980]: I1206 03:35:21.098380 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:21 crc kubenswrapper[4980]: I1206 03:35:21.098392 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:21 crc kubenswrapper[4980]: I1206 03:35:21.098407 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:21 crc kubenswrapper[4980]: I1206 03:35:21.098418 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:21Z","lastTransitionTime":"2025-12-06T03:35:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:21 crc kubenswrapper[4980]: I1206 03:35:21.137015 4980 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-5w4fr_5af0e768-3c9c-4401-ab56-57516cd3170e/ovnkube-controller/2.log" Dec 06 03:35:21 crc kubenswrapper[4980]: I1206 03:35:21.139296 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5w4fr" event={"ID":"5af0e768-3c9c-4401-ab56-57516cd3170e","Type":"ContainerStarted","Data":"33fa9bd73b9af5df6de4e7ed019a512334f465b830a3f4396cde14989658bf55"} Dec 06 03:35:21 crc kubenswrapper[4980]: I1206 03:35:21.139795 4980 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-5w4fr" Dec 06 03:35:21 crc kubenswrapper[4980]: I1206 03:35:21.160012 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7f958150-549d-4c80-8a7e-23c39e4f9500\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c28b6895b14e73f4f3eb7a8ce93076c05092ba9b58ea0e3e879c65bcfa227a21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://76cff8a2d652df38b04eefc053832baf4cbb5db778ff5e525b18c2825141b118\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://206f6f19fe3197f5d6798cc86c5473bf2decb02a2860ee7ffb5efb5f60952e8c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bafaa9670b05671fb57b658c2960bf3ee32c84e52f86a3a3bbab01d5a38ff0c4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://773a0bc2aab73ffc2717e987543defc9b786cc22300aa54b4ba4f3454504e331\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-06T03:34:22Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1206 03:34:15.305330 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1206 03:34:15.307249 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1592846546/tls.crt::/tmp/serving-cert-1592846546/tls.key\\\\\\\"\\\\nI1206 03:34:21.814037 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1206 03:34:21.816756 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1206 03:34:21.816780 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1206 03:34:21.816806 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1206 03:34:21.816813 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1206 03:34:21.831301 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1206 03:34:21.831311 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1206 03:34:21.831324 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 03:34:21.831330 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 03:34:21.831335 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1206 03:34:21.831339 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1206 03:34:21.831342 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1206 03:34:21.831346 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1206 03:34:21.911269 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:04Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d40e0e7446f8416bf9da94fc53c01be937f9a51a11eb101157d2ddd4597720d2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:04Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2cba89f2f1bed96d59f183628d900d9b72c91eb472e273c8bed0babcab485910\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2cba89f2f1bed96d59f183628d900d9b72c91eb472e273c8bed0babcab485910\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:01Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:35:21Z is after 2025-08-24T17:21:41Z" Dec 06 03:35:21 crc kubenswrapper[4980]: I1206 03:35:21.175190 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6ed68f66-1cf6-465a-8f66-76d5af6a44d8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://725e84f6bdfb722a8ce69158d03ea4c66a9f087642d55cb9651e003c81be0c8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://71f7b492ea07cdf7dc65fc48184ec1e64914eb1811bbfdddc0fb14c8ad6cb296\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b543aa4ee03d5a14909b67d876c5675c02e1df2a89a3c9ac8d3fb587fea53d8f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://32af3427e872a70cb3c2357116f04cabc8d3d9b69e7306b2e0de2ae22f9469d4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:01Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:35:21Z is after 2025-08-24T17:21:41Z" Dec 06 03:35:21 crc kubenswrapper[4980]: I1206 03:35:21.183864 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 03:35:21 crc kubenswrapper[4980]: E1206 03:35:21.184061 4980 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 03:35:21 crc kubenswrapper[4980]: I1206 03:35:21.190491 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"64278250-0b19-4591-b06e-6c96652ee92a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5beac3fd93b448506c0a110d70ef5faf42efa18b16abe6b3d4eb7b9c4233ad93\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ae66052702fcfd94f0123f4fbbe9efed37eb04609269a2bc951f95669de16601\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://173903f3554d7ac9158493498a016d618d92f46da1500408834b3c24249cf117\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1f596bfd5321ca225f4be03d4cf229c31a52def9731c0e177ab3a0bc010380e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1f596bfd5321ca225f4be03d4cf229c31a52def9731c0e177ab3a0bc010380e1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:02Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:01Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:35:21Z is after 2025-08-24T17:21:41Z" Dec 06 03:35:21 crc kubenswrapper[4980]: I1206 03:35:21.200302 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:21 crc kubenswrapper[4980]: I1206 03:35:21.200475 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:21 crc kubenswrapper[4980]: I1206 03:35:21.200576 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:21 crc kubenswrapper[4980]: I1206 03:35:21.200648 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:21 crc kubenswrapper[4980]: I1206 03:35:21.200719 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:21Z","lastTransitionTime":"2025-12-06T03:35:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:21 crc kubenswrapper[4980]: I1206 03:35:21.203498 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:22Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:35:21Z is after 2025-08-24T17:21:41Z" Dec 06 03:35:21 crc kubenswrapper[4980]: I1206 03:35:21.218890 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://334ea6a273fd7d268906c9f4a6dbf6f4df4fcdbf38f3d7d248cb7d6cf8537794\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ec44dcd5e4bf0a0466bb586ebd61259364dbbfe22499f4cbff854e3bd153082\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:35:21Z is after 2025-08-24T17:21:41Z" Dec 06 03:35:21 crc kubenswrapper[4980]: I1206 03:35:21.230257 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:22Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:35:21Z is after 2025-08-24T17:21:41Z" Dec 06 03:35:21 crc kubenswrapper[4980]: I1206 03:35:21.241226 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-sv5sz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"873977aa-4eb0-4526-bc0b-757ad2dea2a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01f0cc855c822a71ea8ef6a7ce6e25747ff9956691735973f935c4a28672a7a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lg7tx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:23Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-sv5sz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:35:21Z is after 2025-08-24T17:21:41Z" Dec 06 03:35:21 crc kubenswrapper[4980]: I1206 03:35:21.265431 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5w4fr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5af0e768-3c9c-4401-ab56-57516cd3170e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:23Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:23Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://768b8e1fe1a00a8467d744da5ad3f61006b0a3612c96f8e4d0ee6fcb93195f9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfzrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6342903b318317ee8d84c2f5d467cb25c811a789be0d6ab0e0cba37b5611e16d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfzrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d620685574db7594132ae24a91ec0a93dce20a9a289540deebb39650b8a1816\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfzrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c67ec8f9c1764fcdf215d52822c2d4f0437dc7f056f05d708767121f6c1f3515\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfzrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f35beae5339eeaf26494ed678e3c1bd02349eae2a4b4615325b17923f3b653b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfzrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a940ec6ee523313728f7e8ae934b7e0b0cff82771e3e64723309db1fae013c3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfzrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://33fa9bd73b9af5df6de4e7ed019a512334f465b830a3f4396cde14989658bf55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fae428d1a6eaeb2e6ddd4a5b1a1170e230635739ea0b39743e565e47fb7ee6e4\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-06T03:34:53Z\\\",\\\"message\\\":\\\"/pkg/client/informers/externalversions/factory.go:141\\\\nI1206 03:34:53.908007 6569 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1206 03:34:53.908049 6569 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1206 03:34:53.908094 6569 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1206 03:34:53.908109 6569 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1206 03:34:53.908160 6569 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1206 03:34:53.908208 6569 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1206 03:34:53.908256 6569 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1206 03:34:53.908286 6569 factory.go:656] Stopping watch factory\\\\nI1206 03:34:53.908308 6569 ovnkube.go:599] Stopped ovnkube\\\\nI1206 03:34:53.908349 6569 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1206 03:34:53.908364 6569 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1206 03:34:53.908372 6569 handler.go:208] Removed *v1.Node event handler 2\\\\nI1206 03:34:53.908385 6569 handler.go:208] Removed *v1.Node event handler 7\\\\nI1206 03:34:53.908397 6569 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1206 03:34:53.908415 6569 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1206 03:34:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:52Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:35:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfzrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe261f4322b85702181d7473ab2a9a4d2d4b869d586c3091f811b12a81a63aae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfzrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d876df04b5ab0ba07a5cbcafbe4648403c5014c59eb7d40adee2657c3a255473\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d876df04b5ab0ba07a5cbcafbe4648403c5014c59eb7d40adee2657c3a255473\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfzrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:23Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-5w4fr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:35:21Z is after 2025-08-24T17:21:41Z" Dec 06 03:35:21 crc kubenswrapper[4980]: I1206 03:35:21.286052 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cab608b3-a02f-475b-9aae-9b8f5d878de9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f7fcc44b42bcbb55509048e3a566bb557333c4a83e33d7b0fba8eb514317909\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c22e7dd8886c2dfbb742900ed86ec7b0c2edf9a1363e74f6666961d928f06c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://473e792b2d0a9644854ee0581c6120fc9c989bc342899ef72b1d056080fd4694\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ee8e936dfc39361d6f99460910591ca8c901a20f0d274bbb6aaebfa8a00ee06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://77bc12768b018625ba11a6bfd66688f926c4b7f3bc1d0e41c6551a18b485ecf3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://27cf07cd96893900da4afecd0d56fd9269bc60c1d23246b4685fcd8756357c19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://27cf07cd96893900da4afecd0d56fd9269bc60c1d23246b4685fcd8756357c19\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c49feb24566a8284378670fa0390710d38c237acc94ffb654f90f3fa4626e4c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c49feb24566a8284378670fa0390710d38c237acc94ffb654f90f3fa4626e4c8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:03Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://bb74d1c3d752a379fa394c450217badb45511fbcafad62f7b618131cab94b29e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bb74d1c3d752a379fa394c450217badb45511fbcafad62f7b618131cab94b29e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:01Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:35:21Z is after 2025-08-24T17:21:41Z" Dec 06 03:35:21 crc kubenswrapper[4980]: I1206 03:35:21.302609 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:21 crc kubenswrapper[4980]: I1206 03:35:21.302755 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:21 crc kubenswrapper[4980]: I1206 03:35:21.302849 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:21 crc kubenswrapper[4980]: I1206 03:35:21.302933 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:21 crc kubenswrapper[4980]: I1206 03:35:21.303004 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:21Z","lastTransitionTime":"2025-12-06T03:35:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:21 crc kubenswrapper[4980]: I1206 03:35:21.307453 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://502c727d6baa31f400bee14c34b19abbe38af873f28f6a82a43547431755d981\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:35:21Z is after 2025-08-24T17:21:41Z" Dec 06 03:35:21 crc kubenswrapper[4980]: I1206 03:35:21.333733 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-dx44q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5bac22a-512c-49d0-90db-47de6aaaefd0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://97469dea2cd5393da0acf784ccc1969c8a4d07da7d51b14cdc955013ee9710ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f7mdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3b5dab476c7f227a6e4a2eae8469432667fb9fad345d9419ba59feab9fd385e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c3b5dab476c7f227a6e4a2eae8469432667fb9fad345d9419ba59feab9fd385e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f7mdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e921499231cbdf26b8918c98f9ad8c322fd7848865caeef048470de3e2c0efba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e921499231cbdf26b8918c98f9ad8c322fd7848865caeef048470de3e2c0efba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f7mdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7cbfe7d8cdb89916c238296592fab45b9e92e2c585baa6d67230decd1c84c5bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7cbfe7d8cdb89916c238296592fab45b9e92e2c585baa6d67230decd1c84c5bb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f7mdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f45cbf23b6b0d208cefc4b493f901bab9d71ccf2f1456507d54fb1f30c4fd48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7f45cbf23b6b0d208cefc4b493f901bab9d71ccf2f1456507d54fb1f30c4fd48\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f7mdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://344ce4b145ff021fcaf163803d8ad8cc6e9565bf7f0992e2653c43db663d8411\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://344ce4b145ff021fcaf163803d8ad8cc6e9565bf7f0992e2653c43db663d8411\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f7mdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://619e56a9937dd6376ab471ba16da30e5d77f443f5bd33610bea5ec85cd0446ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://619e56a9937dd6376ab471ba16da30e5d77f443f5bd33610bea5ec85cd0446ef\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f7mdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:23Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-dx44q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:35:21Z is after 2025-08-24T17:21:41Z" Dec 06 03:35:21 crc kubenswrapper[4980]: I1206 03:35:21.344293 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-r5zfb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"320f44d1-a671-4a91-b328-a8b0fdd8f23a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://92393ecaaef8cfcfd01837671f8cc89803dd570b637b4a373b40b7f487788826\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zhm5t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52e8db644463794041978ced2c0591c13e0ada8e5f1a78a94a0a665218a37b5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zhm5t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:23Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-r5zfb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:35:21Z is after 2025-08-24T17:21:41Z" Dec 06 03:35:21 crc kubenswrapper[4980]: I1206 03:35:21.353121 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-cvzg8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f1cde706-268d-4d26-818f-2f42bfb37b5f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://86f9a77acae360740be827eb0d1e1516e8a49cb9b197b6f5f9468ac70a101158\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rk86l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:29Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-cvzg8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:35:21Z is after 2025-08-24T17:21:41Z" Dec 06 03:35:21 crc kubenswrapper[4980]: I1206 03:35:21.363891 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:35:21Z is after 2025-08-24T17:21:41Z" Dec 06 03:35:21 crc kubenswrapper[4980]: I1206 03:35:21.375642 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4ef01e7d9809fd9cfc46e3a58b07159cdc6f5a3c210d1c9efc49bca27b30858\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:35:21Z is after 2025-08-24T17:21:41Z" Dec 06 03:35:21 crc kubenswrapper[4980]: I1206 03:35:21.386891 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-kq4z8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f96c1f6c-55e8-4d86-bff3-5381581b7022\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:35:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:35:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ea7f0bc23247b5effe7f0250582c8694bbf0e9a2d185a844773665a0a48f350e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://799498681afa7653408bc23b1e82d75c16e3c809d2ee7f322780e20c537eb710\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-06T03:35:15Z\\\",\\\"message\\\":\\\"2025-12-06T03:34:30+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_68ae7793-ff25-4e9c-8158-bea74df4ae16\\\\n2025-12-06T03:34:30+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_68ae7793-ff25-4e9c-8158-bea74df4ae16 to /host/opt/cni/bin/\\\\n2025-12-06T03:34:30Z [verbose] multus-daemon started\\\\n2025-12-06T03:34:30Z [verbose] Readiness Indicator file check\\\\n2025-12-06T03:35:15Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:24Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:35:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bdx7g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:23Z\\\"}}\" for pod \"openshift-multus\"/\"multus-kq4z8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:35:21Z is after 2025-08-24T17:21:41Z" Dec 06 03:35:21 crc kubenswrapper[4980]: I1206 03:35:21.396614 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-6kxfq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"951b6deb-bcea-4332-8ae8-c23ac7aaef36\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://423d30c3e65b04bdf6ff5ef53d8a7575a434c05e28489370a7370381ebc29de9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b8htf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://282df50dd9653c1cbe97fc2041d9e9ea91a992661682652aafc12e566d8f1b81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b8htf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:37Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-6kxfq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:35:21Z is after 2025-08-24T17:21:41Z" Dec 06 03:35:21 crc kubenswrapper[4980]: I1206 03:35:21.405404 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:21 crc kubenswrapper[4980]: I1206 03:35:21.405446 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:21 crc kubenswrapper[4980]: I1206 03:35:21.405454 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:21 crc kubenswrapper[4980]: I1206 03:35:21.405478 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:21 crc kubenswrapper[4980]: I1206 03:35:21.405490 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:21Z","lastTransitionTime":"2025-12-06T03:35:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:21 crc kubenswrapper[4980]: I1206 03:35:21.407067 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-mcprh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e6e88c5e-a255-449a-ae38-ac4d73a8e984\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g7xz9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g7xz9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:38Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-mcprh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:35:21Z is after 2025-08-24T17:21:41Z" Dec 06 03:35:21 crc kubenswrapper[4980]: I1206 03:35:21.418893 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7f958150-549d-4c80-8a7e-23c39e4f9500\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c28b6895b14e73f4f3eb7a8ce93076c05092ba9b58ea0e3e879c65bcfa227a21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://76cff8a2d652df38b04eefc053832baf4cbb5db778ff5e525b18c2825141b118\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://206f6f19fe3197f5d6798cc86c5473bf2decb02a2860ee7ffb5efb5f60952e8c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bafaa9670b05671fb57b658c2960bf3ee32c84e52f86a3a3bbab01d5a38ff0c4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://773a0bc2aab73ffc2717e987543defc9b786cc22300aa54b4ba4f3454504e331\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-06T03:34:22Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1206 03:34:15.305330 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1206 03:34:15.307249 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1592846546/tls.crt::/tmp/serving-cert-1592846546/tls.key\\\\\\\"\\\\nI1206 03:34:21.814037 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1206 03:34:21.816756 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1206 03:34:21.816780 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1206 03:34:21.816806 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1206 03:34:21.816813 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1206 03:34:21.831301 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1206 03:34:21.831311 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1206 03:34:21.831324 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 03:34:21.831330 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 03:34:21.831335 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1206 03:34:21.831339 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1206 03:34:21.831342 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1206 03:34:21.831346 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1206 03:34:21.911269 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:04Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d40e0e7446f8416bf9da94fc53c01be937f9a51a11eb101157d2ddd4597720d2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:04Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2cba89f2f1bed96d59f183628d900d9b72c91eb472e273c8bed0babcab485910\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2cba89f2f1bed96d59f183628d900d9b72c91eb472e273c8bed0babcab485910\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:01Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:35:21Z is after 2025-08-24T17:21:41Z" Dec 06 03:35:21 crc kubenswrapper[4980]: I1206 03:35:21.432820 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6ed68f66-1cf6-465a-8f66-76d5af6a44d8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://725e84f6bdfb722a8ce69158d03ea4c66a9f087642d55cb9651e003c81be0c8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://71f7b492ea07cdf7dc65fc48184ec1e64914eb1811bbfdddc0fb14c8ad6cb296\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b543aa4ee03d5a14909b67d876c5675c02e1df2a89a3c9ac8d3fb587fea53d8f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://32af3427e872a70cb3c2357116f04cabc8d3d9b69e7306b2e0de2ae22f9469d4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:01Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:35:21Z is after 2025-08-24T17:21:41Z" Dec 06 03:35:21 crc kubenswrapper[4980]: I1206 03:35:21.446584 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"64278250-0b19-4591-b06e-6c96652ee92a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5beac3fd93b448506c0a110d70ef5faf42efa18b16abe6b3d4eb7b9c4233ad93\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ae66052702fcfd94f0123f4fbbe9efed37eb04609269a2bc951f95669de16601\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://173903f3554d7ac9158493498a016d618d92f46da1500408834b3c24249cf117\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1f596bfd5321ca225f4be03d4cf229c31a52def9731c0e177ab3a0bc010380e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1f596bfd5321ca225f4be03d4cf229c31a52def9731c0e177ab3a0bc010380e1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:02Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:01Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:35:21Z is after 2025-08-24T17:21:41Z" Dec 06 03:35:21 crc kubenswrapper[4980]: I1206 03:35:21.463581 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:22Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:35:21Z is after 2025-08-24T17:21:41Z" Dec 06 03:35:21 crc kubenswrapper[4980]: I1206 03:35:21.482111 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://334ea6a273fd7d268906c9f4a6dbf6f4df4fcdbf38f3d7d248cb7d6cf8537794\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ec44dcd5e4bf0a0466bb586ebd61259364dbbfe22499f4cbff854e3bd153082\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:35:21Z is after 2025-08-24T17:21:41Z" Dec 06 03:35:21 crc kubenswrapper[4980]: I1206 03:35:21.502361 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:22Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:35:21Z is after 2025-08-24T17:21:41Z" Dec 06 03:35:21 crc kubenswrapper[4980]: I1206 03:35:21.508264 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:21 crc kubenswrapper[4980]: I1206 03:35:21.508424 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:21 crc kubenswrapper[4980]: I1206 03:35:21.508454 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:21 crc kubenswrapper[4980]: I1206 03:35:21.508489 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:21 crc kubenswrapper[4980]: I1206 03:35:21.508583 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:21Z","lastTransitionTime":"2025-12-06T03:35:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:21 crc kubenswrapper[4980]: I1206 03:35:21.523238 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-sv5sz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"873977aa-4eb0-4526-bc0b-757ad2dea2a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01f0cc855c822a71ea8ef6a7ce6e25747ff9956691735973f935c4a28672a7a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lg7tx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:23Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-sv5sz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:35:21Z is after 2025-08-24T17:21:41Z" Dec 06 03:35:21 crc kubenswrapper[4980]: I1206 03:35:21.545958 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5w4fr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5af0e768-3c9c-4401-ab56-57516cd3170e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:23Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:23Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://768b8e1fe1a00a8467d744da5ad3f61006b0a3612c96f8e4d0ee6fcb93195f9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfzrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6342903b318317ee8d84c2f5d467cb25c811a789be0d6ab0e0cba37b5611e16d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfzrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d620685574db7594132ae24a91ec0a93dce20a9a289540deebb39650b8a1816\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfzrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c67ec8f9c1764fcdf215d52822c2d4f0437dc7f056f05d708767121f6c1f3515\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfzrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f35beae5339eeaf26494ed678e3c1bd02349eae2a4b4615325b17923f3b653b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfzrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a940ec6ee523313728f7e8ae934b7e0b0cff82771e3e64723309db1fae013c3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfzrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://33fa9bd73b9af5df6de4e7ed019a512334f465b830a3f4396cde14989658bf55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fae428d1a6eaeb2e6ddd4a5b1a1170e230635739ea0b39743e565e47fb7ee6e4\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-06T03:34:53Z\\\",\\\"message\\\":\\\"/pkg/client/informers/externalversions/factory.go:141\\\\nI1206 03:34:53.908007 6569 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1206 03:34:53.908049 6569 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1206 03:34:53.908094 6569 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1206 03:34:53.908109 6569 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1206 03:34:53.908160 6569 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1206 03:34:53.908208 6569 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1206 03:34:53.908256 6569 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1206 03:34:53.908286 6569 factory.go:656] Stopping watch factory\\\\nI1206 03:34:53.908308 6569 ovnkube.go:599] Stopped ovnkube\\\\nI1206 03:34:53.908349 6569 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1206 03:34:53.908364 6569 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1206 03:34:53.908372 6569 handler.go:208] Removed *v1.Node event handler 2\\\\nI1206 03:34:53.908385 6569 handler.go:208] Removed *v1.Node event handler 7\\\\nI1206 03:34:53.908397 6569 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1206 03:34:53.908415 6569 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1206 03:34:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:52Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:35:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfzrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe261f4322b85702181d7473ab2a9a4d2d4b869d586c3091f811b12a81a63aae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfzrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d876df04b5ab0ba07a5cbcafbe4648403c5014c59eb7d40adee2657c3a255473\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d876df04b5ab0ba07a5cbcafbe4648403c5014c59eb7d40adee2657c3a255473\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfzrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:23Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-5w4fr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:35:21Z is after 2025-08-24T17:21:41Z" Dec 06 03:35:21 crc kubenswrapper[4980]: I1206 03:35:21.573554 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cab608b3-a02f-475b-9aae-9b8f5d878de9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f7fcc44b42bcbb55509048e3a566bb557333c4a83e33d7b0fba8eb514317909\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c22e7dd8886c2dfbb742900ed86ec7b0c2edf9a1363e74f6666961d928f06c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://473e792b2d0a9644854ee0581c6120fc9c989bc342899ef72b1d056080fd4694\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ee8e936dfc39361d6f99460910591ca8c901a20f0d274bbb6aaebfa8a00ee06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://77bc12768b018625ba11a6bfd66688f926c4b7f3bc1d0e41c6551a18b485ecf3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://27cf07cd96893900da4afecd0d56fd9269bc60c1d23246b4685fcd8756357c19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://27cf07cd96893900da4afecd0d56fd9269bc60c1d23246b4685fcd8756357c19\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c49feb24566a8284378670fa0390710d38c237acc94ffb654f90f3fa4626e4c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c49feb24566a8284378670fa0390710d38c237acc94ffb654f90f3fa4626e4c8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:03Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://bb74d1c3d752a379fa394c450217badb45511fbcafad62f7b618131cab94b29e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bb74d1c3d752a379fa394c450217badb45511fbcafad62f7b618131cab94b29e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:01Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:35:21Z is after 2025-08-24T17:21:41Z" Dec 06 03:35:21 crc kubenswrapper[4980]: I1206 03:35:21.587266 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://502c727d6baa31f400bee14c34b19abbe38af873f28f6a82a43547431755d981\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:35:21Z is after 2025-08-24T17:21:41Z" Dec 06 03:35:21 crc kubenswrapper[4980]: I1206 03:35:21.605298 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-dx44q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5bac22a-512c-49d0-90db-47de6aaaefd0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://97469dea2cd5393da0acf784ccc1969c8a4d07da7d51b14cdc955013ee9710ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f7mdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3b5dab476c7f227a6e4a2eae8469432667fb9fad345d9419ba59feab9fd385e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c3b5dab476c7f227a6e4a2eae8469432667fb9fad345d9419ba59feab9fd385e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f7mdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e921499231cbdf26b8918c98f9ad8c322fd7848865caeef048470de3e2c0efba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e921499231cbdf26b8918c98f9ad8c322fd7848865caeef048470de3e2c0efba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f7mdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7cbfe7d8cdb89916c238296592fab45b9e92e2c585baa6d67230decd1c84c5bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7cbfe7d8cdb89916c238296592fab45b9e92e2c585baa6d67230decd1c84c5bb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f7mdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f45cbf23b6b0d208cefc4b493f901bab9d71ccf2f1456507d54fb1f30c4fd48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7f45cbf23b6b0d208cefc4b493f901bab9d71ccf2f1456507d54fb1f30c4fd48\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f7mdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://344ce4b145ff021fcaf163803d8ad8cc6e9565bf7f0992e2653c43db663d8411\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://344ce4b145ff021fcaf163803d8ad8cc6e9565bf7f0992e2653c43db663d8411\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f7mdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://619e56a9937dd6376ab471ba16da30e5d77f443f5bd33610bea5ec85cd0446ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://619e56a9937dd6376ab471ba16da30e5d77f443f5bd33610bea5ec85cd0446ef\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f7mdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:23Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-dx44q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:35:21Z is after 2025-08-24T17:21:41Z" Dec 06 03:35:21 crc kubenswrapper[4980]: I1206 03:35:21.611616 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:21 crc kubenswrapper[4980]: I1206 03:35:21.611679 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:21 crc kubenswrapper[4980]: I1206 03:35:21.611696 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:21 crc kubenswrapper[4980]: I1206 03:35:21.611724 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:21 crc kubenswrapper[4980]: I1206 03:35:21.611739 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:21Z","lastTransitionTime":"2025-12-06T03:35:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:21 crc kubenswrapper[4980]: I1206 03:35:21.618470 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-r5zfb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"320f44d1-a671-4a91-b328-a8b0fdd8f23a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://92393ecaaef8cfcfd01837671f8cc89803dd570b637b4a373b40b7f487788826\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zhm5t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52e8db644463794041978ced2c0591c13e0ada8e5f1a78a94a0a665218a37b5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zhm5t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:23Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-r5zfb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:35:21Z is after 2025-08-24T17:21:41Z" Dec 06 03:35:21 crc kubenswrapper[4980]: I1206 03:35:21.628523 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-cvzg8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f1cde706-268d-4d26-818f-2f42bfb37b5f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://86f9a77acae360740be827eb0d1e1516e8a49cb9b197b6f5f9468ac70a101158\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rk86l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:29Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-cvzg8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:35:21Z is after 2025-08-24T17:21:41Z" Dec 06 03:35:21 crc kubenswrapper[4980]: I1206 03:35:21.643337 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:35:21Z is after 2025-08-24T17:21:41Z" Dec 06 03:35:21 crc kubenswrapper[4980]: I1206 03:35:21.657484 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4ef01e7d9809fd9cfc46e3a58b07159cdc6f5a3c210d1c9efc49bca27b30858\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:35:21Z is after 2025-08-24T17:21:41Z" Dec 06 03:35:21 crc kubenswrapper[4980]: I1206 03:35:21.675931 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-kq4z8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f96c1f6c-55e8-4d86-bff3-5381581b7022\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:35:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:35:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ea7f0bc23247b5effe7f0250582c8694bbf0e9a2d185a844773665a0a48f350e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://799498681afa7653408bc23b1e82d75c16e3c809d2ee7f322780e20c537eb710\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-06T03:35:15Z\\\",\\\"message\\\":\\\"2025-12-06T03:34:30+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_68ae7793-ff25-4e9c-8158-bea74df4ae16\\\\n2025-12-06T03:34:30+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_68ae7793-ff25-4e9c-8158-bea74df4ae16 to /host/opt/cni/bin/\\\\n2025-12-06T03:34:30Z [verbose] multus-daemon started\\\\n2025-12-06T03:34:30Z [verbose] Readiness Indicator file check\\\\n2025-12-06T03:35:15Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:24Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:35:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bdx7g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:23Z\\\"}}\" for pod \"openshift-multus\"/\"multus-kq4z8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:35:21Z is after 2025-08-24T17:21:41Z" Dec 06 03:35:21 crc kubenswrapper[4980]: I1206 03:35:21.687160 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-6kxfq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"951b6deb-bcea-4332-8ae8-c23ac7aaef36\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://423d30c3e65b04bdf6ff5ef53d8a7575a434c05e28489370a7370381ebc29de9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b8htf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://282df50dd9653c1cbe97fc2041d9e9ea91a992661682652aafc12e566d8f1b81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b8htf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:37Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-6kxfq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:35:21Z is after 2025-08-24T17:21:41Z" Dec 06 03:35:21 crc kubenswrapper[4980]: I1206 03:35:21.698864 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-mcprh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e6e88c5e-a255-449a-ae38-ac4d73a8e984\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g7xz9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g7xz9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:38Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-mcprh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:35:21Z is after 2025-08-24T17:21:41Z" Dec 06 03:35:21 crc kubenswrapper[4980]: I1206 03:35:21.714069 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:21 crc kubenswrapper[4980]: I1206 03:35:21.714122 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:21 crc kubenswrapper[4980]: I1206 03:35:21.714138 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:21 crc kubenswrapper[4980]: I1206 03:35:21.714158 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:21 crc kubenswrapper[4980]: I1206 03:35:21.714172 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:21Z","lastTransitionTime":"2025-12-06T03:35:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:21 crc kubenswrapper[4980]: I1206 03:35:21.817894 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:21 crc kubenswrapper[4980]: I1206 03:35:21.817957 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:21 crc kubenswrapper[4980]: I1206 03:35:21.817974 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:21 crc kubenswrapper[4980]: I1206 03:35:21.818000 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:21 crc kubenswrapper[4980]: I1206 03:35:21.818024 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:21Z","lastTransitionTime":"2025-12-06T03:35:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:21 crc kubenswrapper[4980]: I1206 03:35:21.920297 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:21 crc kubenswrapper[4980]: I1206 03:35:21.920340 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:21 crc kubenswrapper[4980]: I1206 03:35:21.920352 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:21 crc kubenswrapper[4980]: I1206 03:35:21.920370 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:21 crc kubenswrapper[4980]: I1206 03:35:21.920382 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:21Z","lastTransitionTime":"2025-12-06T03:35:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:22 crc kubenswrapper[4980]: I1206 03:35:22.022950 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:22 crc kubenswrapper[4980]: I1206 03:35:22.023010 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:22 crc kubenswrapper[4980]: I1206 03:35:22.023027 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:22 crc kubenswrapper[4980]: I1206 03:35:22.023049 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:22 crc kubenswrapper[4980]: I1206 03:35:22.023062 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:22Z","lastTransitionTime":"2025-12-06T03:35:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:22 crc kubenswrapper[4980]: I1206 03:35:22.125402 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:22 crc kubenswrapper[4980]: I1206 03:35:22.125442 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:22 crc kubenswrapper[4980]: I1206 03:35:22.125451 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:22 crc kubenswrapper[4980]: I1206 03:35:22.125463 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:22 crc kubenswrapper[4980]: I1206 03:35:22.125472 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:22Z","lastTransitionTime":"2025-12-06T03:35:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:22 crc kubenswrapper[4980]: I1206 03:35:22.145362 4980 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-5w4fr_5af0e768-3c9c-4401-ab56-57516cd3170e/ovnkube-controller/3.log" Dec 06 03:35:22 crc kubenswrapper[4980]: I1206 03:35:22.145920 4980 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-5w4fr_5af0e768-3c9c-4401-ab56-57516cd3170e/ovnkube-controller/2.log" Dec 06 03:35:22 crc kubenswrapper[4980]: I1206 03:35:22.149068 4980 generic.go:334] "Generic (PLEG): container finished" podID="5af0e768-3c9c-4401-ab56-57516cd3170e" containerID="33fa9bd73b9af5df6de4e7ed019a512334f465b830a3f4396cde14989658bf55" exitCode=1 Dec 06 03:35:22 crc kubenswrapper[4980]: I1206 03:35:22.149129 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5w4fr" event={"ID":"5af0e768-3c9c-4401-ab56-57516cd3170e","Type":"ContainerDied","Data":"33fa9bd73b9af5df6de4e7ed019a512334f465b830a3f4396cde14989658bf55"} Dec 06 03:35:22 crc kubenswrapper[4980]: I1206 03:35:22.149174 4980 scope.go:117] "RemoveContainer" containerID="fae428d1a6eaeb2e6ddd4a5b1a1170e230635739ea0b39743e565e47fb7ee6e4" Dec 06 03:35:22 crc kubenswrapper[4980]: I1206 03:35:22.150045 4980 scope.go:117] "RemoveContainer" containerID="33fa9bd73b9af5df6de4e7ed019a512334f465b830a3f4396cde14989658bf55" Dec 06 03:35:22 crc kubenswrapper[4980]: E1206 03:35:22.150244 4980 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-5w4fr_openshift-ovn-kubernetes(5af0e768-3c9c-4401-ab56-57516cd3170e)\"" pod="openshift-ovn-kubernetes/ovnkube-node-5w4fr" podUID="5af0e768-3c9c-4401-ab56-57516cd3170e" Dec 06 03:35:22 crc kubenswrapper[4980]: I1206 03:35:22.161196 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"64278250-0b19-4591-b06e-6c96652ee92a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5beac3fd93b448506c0a110d70ef5faf42efa18b16abe6b3d4eb7b9c4233ad93\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ae66052702fcfd94f0123f4fbbe9efed37eb04609269a2bc951f95669de16601\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://173903f3554d7ac9158493498a016d618d92f46da1500408834b3c24249cf117\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1f596bfd5321ca225f4be03d4cf229c31a52def9731c0e177ab3a0bc010380e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1f596bfd5321ca225f4be03d4cf229c31a52def9731c0e177ab3a0bc010380e1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:02Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:01Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:35:22Z is after 2025-08-24T17:21:41Z" Dec 06 03:35:22 crc kubenswrapper[4980]: I1206 03:35:22.175447 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:22Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:35:22Z is after 2025-08-24T17:21:41Z" Dec 06 03:35:22 crc kubenswrapper[4980]: I1206 03:35:22.183531 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 03:35:22 crc kubenswrapper[4980]: I1206 03:35:22.183565 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-mcprh" Dec 06 03:35:22 crc kubenswrapper[4980]: E1206 03:35:22.183649 4980 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 03:35:22 crc kubenswrapper[4980]: E1206 03:35:22.183779 4980 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-mcprh" podUID="e6e88c5e-a255-449a-ae38-ac4d73a8e984" Dec 06 03:35:22 crc kubenswrapper[4980]: I1206 03:35:22.183804 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 03:35:22 crc kubenswrapper[4980]: E1206 03:35:22.183850 4980 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 03:35:22 crc kubenswrapper[4980]: I1206 03:35:22.186296 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://334ea6a273fd7d268906c9f4a6dbf6f4df4fcdbf38f3d7d248cb7d6cf8537794\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ec44dcd5e4bf0a0466bb586ebd61259364dbbfe22499f4cbff854e3bd153082\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:35:22Z is after 2025-08-24T17:21:41Z" Dec 06 03:35:22 crc kubenswrapper[4980]: I1206 03:35:22.194844 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:22Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:35:22Z is after 2025-08-24T17:21:41Z" Dec 06 03:35:22 crc kubenswrapper[4980]: I1206 03:35:22.201707 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-sv5sz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"873977aa-4eb0-4526-bc0b-757ad2dea2a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01f0cc855c822a71ea8ef6a7ce6e25747ff9956691735973f935c4a28672a7a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lg7tx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:23Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-sv5sz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:35:22Z is after 2025-08-24T17:21:41Z" Dec 06 03:35:22 crc kubenswrapper[4980]: I1206 03:35:22.216601 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5w4fr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5af0e768-3c9c-4401-ab56-57516cd3170e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:23Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:23Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://768b8e1fe1a00a8467d744da5ad3f61006b0a3612c96f8e4d0ee6fcb93195f9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfzrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6342903b318317ee8d84c2f5d467cb25c811a789be0d6ab0e0cba37b5611e16d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfzrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d620685574db7594132ae24a91ec0a93dce20a9a289540deebb39650b8a1816\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfzrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c67ec8f9c1764fcdf215d52822c2d4f0437dc7f056f05d708767121f6c1f3515\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfzrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f35beae5339eeaf26494ed678e3c1bd02349eae2a4b4615325b17923f3b653b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfzrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a940ec6ee523313728f7e8ae934b7e0b0cff82771e3e64723309db1fae013c3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfzrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://33fa9bd73b9af5df6de4e7ed019a512334f465b830a3f4396cde14989658bf55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fae428d1a6eaeb2e6ddd4a5b1a1170e230635739ea0b39743e565e47fb7ee6e4\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-06T03:34:53Z\\\",\\\"message\\\":\\\"/pkg/client/informers/externalversions/factory.go:141\\\\nI1206 03:34:53.908007 6569 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1206 03:34:53.908049 6569 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1206 03:34:53.908094 6569 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1206 03:34:53.908109 6569 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1206 03:34:53.908160 6569 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1206 03:34:53.908208 6569 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1206 03:34:53.908256 6569 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1206 03:34:53.908286 6569 factory.go:656] Stopping watch factory\\\\nI1206 03:34:53.908308 6569 ovnkube.go:599] Stopped ovnkube\\\\nI1206 03:34:53.908349 6569 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1206 03:34:53.908364 6569 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1206 03:34:53.908372 6569 handler.go:208] Removed *v1.Node event handler 2\\\\nI1206 03:34:53.908385 6569 handler.go:208] Removed *v1.Node event handler 7\\\\nI1206 03:34:53.908397 6569 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1206 03:34:53.908415 6569 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1206 03:34:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:52Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://33fa9bd73b9af5df6de4e7ed019a512334f465b830a3f4396cde14989658bf55\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-06T03:35:21Z\\\",\\\"message\\\":\\\"oller.go:360] Finished syncing service ingress-canary on namespace openshift-ingress-canary for network=default : 580.456µs\\\\nF1206 03:35:21.247527 6909 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:35:21Z is after 2025-08-24T17:21:41Z]\\\\nI1206 03:35:21.247533 6909 obj_retry.go:303] Retry object setup: *v1.Pod openshift-machine-config-operator/machine-config-daemon-r5zfb\\\\nI1206 03:35:21.247536 6909 obj_retry.go:303] Retry object setup: *v1.Pod openshift-network-operator/network-operator-58b4c7f79c-55gtf\\\\nI1206 03:35:21.247529 6909 obj_retry.go:303] Retry object setup: *v1.Pod openshift-network-console/networking-con\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T03:35:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfzrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe261f4322b85702181d7473ab2a9a4d2d4b869d586c3091f811b12a81a63aae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfzrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d876df04b5ab0ba07a5cbcafbe4648403c5014c59eb7d40adee2657c3a255473\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d876df04b5ab0ba07a5cbcafbe4648403c5014c59eb7d40adee2657c3a255473\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfzrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:23Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-5w4fr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:35:22Z is after 2025-08-24T17:21:41Z" Dec 06 03:35:22 crc kubenswrapper[4980]: I1206 03:35:22.228432 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:22 crc kubenswrapper[4980]: I1206 03:35:22.228479 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:22 crc kubenswrapper[4980]: I1206 03:35:22.228494 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:22 crc kubenswrapper[4980]: I1206 03:35:22.228534 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:22 crc kubenswrapper[4980]: I1206 03:35:22.228552 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:22Z","lastTransitionTime":"2025-12-06T03:35:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:22 crc kubenswrapper[4980]: I1206 03:35:22.241181 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cab608b3-a02f-475b-9aae-9b8f5d878de9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f7fcc44b42bcbb55509048e3a566bb557333c4a83e33d7b0fba8eb514317909\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c22e7dd8886c2dfbb742900ed86ec7b0c2edf9a1363e74f6666961d928f06c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://473e792b2d0a9644854ee0581c6120fc9c989bc342899ef72b1d056080fd4694\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ee8e936dfc39361d6f99460910591ca8c901a20f0d274bbb6aaebfa8a00ee06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://77bc12768b018625ba11a6bfd66688f926c4b7f3bc1d0e41c6551a18b485ecf3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://27cf07cd96893900da4afecd0d56fd9269bc60c1d23246b4685fcd8756357c19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://27cf07cd96893900da4afecd0d56fd9269bc60c1d23246b4685fcd8756357c19\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c49feb24566a8284378670fa0390710d38c237acc94ffb654f90f3fa4626e4c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c49feb24566a8284378670fa0390710d38c237acc94ffb654f90f3fa4626e4c8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:03Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://bb74d1c3d752a379fa394c450217badb45511fbcafad62f7b618131cab94b29e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bb74d1c3d752a379fa394c450217badb45511fbcafad62f7b618131cab94b29e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:01Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:35:22Z is after 2025-08-24T17:21:41Z" Dec 06 03:35:22 crc kubenswrapper[4980]: I1206 03:35:22.251560 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://502c727d6baa31f400bee14c34b19abbe38af873f28f6a82a43547431755d981\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:35:22Z is after 2025-08-24T17:21:41Z" Dec 06 03:35:22 crc kubenswrapper[4980]: I1206 03:35:22.263638 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-dx44q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5bac22a-512c-49d0-90db-47de6aaaefd0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://97469dea2cd5393da0acf784ccc1969c8a4d07da7d51b14cdc955013ee9710ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f7mdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3b5dab476c7f227a6e4a2eae8469432667fb9fad345d9419ba59feab9fd385e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c3b5dab476c7f227a6e4a2eae8469432667fb9fad345d9419ba59feab9fd385e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f7mdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e921499231cbdf26b8918c98f9ad8c322fd7848865caeef048470de3e2c0efba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e921499231cbdf26b8918c98f9ad8c322fd7848865caeef048470de3e2c0efba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f7mdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7cbfe7d8cdb89916c238296592fab45b9e92e2c585baa6d67230decd1c84c5bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7cbfe7d8cdb89916c238296592fab45b9e92e2c585baa6d67230decd1c84c5bb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f7mdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f45cbf23b6b0d208cefc4b493f901bab9d71ccf2f1456507d54fb1f30c4fd48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7f45cbf23b6b0d208cefc4b493f901bab9d71ccf2f1456507d54fb1f30c4fd48\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f7mdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://344ce4b145ff021fcaf163803d8ad8cc6e9565bf7f0992e2653c43db663d8411\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://344ce4b145ff021fcaf163803d8ad8cc6e9565bf7f0992e2653c43db663d8411\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f7mdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://619e56a9937dd6376ab471ba16da30e5d77f443f5bd33610bea5ec85cd0446ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://619e56a9937dd6376ab471ba16da30e5d77f443f5bd33610bea5ec85cd0446ef\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f7mdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:23Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-dx44q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:35:22Z is after 2025-08-24T17:21:41Z" Dec 06 03:35:22 crc kubenswrapper[4980]: I1206 03:35:22.278580 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-r5zfb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"320f44d1-a671-4a91-b328-a8b0fdd8f23a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://92393ecaaef8cfcfd01837671f8cc89803dd570b637b4a373b40b7f487788826\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zhm5t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52e8db644463794041978ced2c0591c13e0ada8e5f1a78a94a0a665218a37b5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zhm5t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:23Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-r5zfb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:35:22Z is after 2025-08-24T17:21:41Z" Dec 06 03:35:22 crc kubenswrapper[4980]: I1206 03:35:22.289795 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-cvzg8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f1cde706-268d-4d26-818f-2f42bfb37b5f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://86f9a77acae360740be827eb0d1e1516e8a49cb9b197b6f5f9468ac70a101158\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rk86l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:29Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-cvzg8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:35:22Z is after 2025-08-24T17:21:41Z" Dec 06 03:35:22 crc kubenswrapper[4980]: I1206 03:35:22.301679 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:35:22Z is after 2025-08-24T17:21:41Z" Dec 06 03:35:22 crc kubenswrapper[4980]: I1206 03:35:22.315893 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4ef01e7d9809fd9cfc46e3a58b07159cdc6f5a3c210d1c9efc49bca27b30858\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:35:22Z is after 2025-08-24T17:21:41Z" Dec 06 03:35:22 crc kubenswrapper[4980]: I1206 03:35:22.328290 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-kq4z8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f96c1f6c-55e8-4d86-bff3-5381581b7022\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:35:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:35:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ea7f0bc23247b5effe7f0250582c8694bbf0e9a2d185a844773665a0a48f350e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://799498681afa7653408bc23b1e82d75c16e3c809d2ee7f322780e20c537eb710\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-06T03:35:15Z\\\",\\\"message\\\":\\\"2025-12-06T03:34:30+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_68ae7793-ff25-4e9c-8158-bea74df4ae16\\\\n2025-12-06T03:34:30+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_68ae7793-ff25-4e9c-8158-bea74df4ae16 to /host/opt/cni/bin/\\\\n2025-12-06T03:34:30Z [verbose] multus-daemon started\\\\n2025-12-06T03:34:30Z [verbose] Readiness Indicator file check\\\\n2025-12-06T03:35:15Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:24Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:35:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bdx7g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:23Z\\\"}}\" for pod \"openshift-multus\"/\"multus-kq4z8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:35:22Z is after 2025-08-24T17:21:41Z" Dec 06 03:35:22 crc kubenswrapper[4980]: I1206 03:35:22.331444 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:22 crc kubenswrapper[4980]: I1206 03:35:22.331475 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:22 crc kubenswrapper[4980]: I1206 03:35:22.331483 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:22 crc kubenswrapper[4980]: I1206 03:35:22.331495 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:22 crc kubenswrapper[4980]: I1206 03:35:22.331506 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:22Z","lastTransitionTime":"2025-12-06T03:35:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:22 crc kubenswrapper[4980]: I1206 03:35:22.342595 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-6kxfq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"951b6deb-bcea-4332-8ae8-c23ac7aaef36\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://423d30c3e65b04bdf6ff5ef53d8a7575a434c05e28489370a7370381ebc29de9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b8htf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://282df50dd9653c1cbe97fc2041d9e9ea91a992661682652aafc12e566d8f1b81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b8htf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:37Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-6kxfq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:35:22Z is after 2025-08-24T17:21:41Z" Dec 06 03:35:22 crc kubenswrapper[4980]: I1206 03:35:22.352799 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-mcprh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e6e88c5e-a255-449a-ae38-ac4d73a8e984\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g7xz9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g7xz9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:38Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-mcprh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:35:22Z is after 2025-08-24T17:21:41Z" Dec 06 03:35:22 crc kubenswrapper[4980]: I1206 03:35:22.364558 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7f958150-549d-4c80-8a7e-23c39e4f9500\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c28b6895b14e73f4f3eb7a8ce93076c05092ba9b58ea0e3e879c65bcfa227a21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://76cff8a2d652df38b04eefc053832baf4cbb5db778ff5e525b18c2825141b118\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://206f6f19fe3197f5d6798cc86c5473bf2decb02a2860ee7ffb5efb5f60952e8c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bafaa9670b05671fb57b658c2960bf3ee32c84e52f86a3a3bbab01d5a38ff0c4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://773a0bc2aab73ffc2717e987543defc9b786cc22300aa54b4ba4f3454504e331\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-06T03:34:22Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1206 03:34:15.305330 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1206 03:34:15.307249 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1592846546/tls.crt::/tmp/serving-cert-1592846546/tls.key\\\\\\\"\\\\nI1206 03:34:21.814037 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1206 03:34:21.816756 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1206 03:34:21.816780 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1206 03:34:21.816806 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1206 03:34:21.816813 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1206 03:34:21.831301 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1206 03:34:21.831311 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1206 03:34:21.831324 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 03:34:21.831330 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 03:34:21.831335 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1206 03:34:21.831339 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1206 03:34:21.831342 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1206 03:34:21.831346 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1206 03:34:21.911269 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:04Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d40e0e7446f8416bf9da94fc53c01be937f9a51a11eb101157d2ddd4597720d2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:04Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2cba89f2f1bed96d59f183628d900d9b72c91eb472e273c8bed0babcab485910\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2cba89f2f1bed96d59f183628d900d9b72c91eb472e273c8bed0babcab485910\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:01Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:35:22Z is after 2025-08-24T17:21:41Z" Dec 06 03:35:22 crc kubenswrapper[4980]: I1206 03:35:22.378460 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6ed68f66-1cf6-465a-8f66-76d5af6a44d8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://725e84f6bdfb722a8ce69158d03ea4c66a9f087642d55cb9651e003c81be0c8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://71f7b492ea07cdf7dc65fc48184ec1e64914eb1811bbfdddc0fb14c8ad6cb296\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b543aa4ee03d5a14909b67d876c5675c02e1df2a89a3c9ac8d3fb587fea53d8f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://32af3427e872a70cb3c2357116f04cabc8d3d9b69e7306b2e0de2ae22f9469d4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:01Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:35:22Z is after 2025-08-24T17:21:41Z" Dec 06 03:35:22 crc kubenswrapper[4980]: I1206 03:35:22.433629 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:22 crc kubenswrapper[4980]: I1206 03:35:22.433669 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:22 crc kubenswrapper[4980]: I1206 03:35:22.433680 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:22 crc kubenswrapper[4980]: I1206 03:35:22.433694 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:22 crc kubenswrapper[4980]: I1206 03:35:22.433703 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:22Z","lastTransitionTime":"2025-12-06T03:35:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:22 crc kubenswrapper[4980]: I1206 03:35:22.536028 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:22 crc kubenswrapper[4980]: I1206 03:35:22.536089 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:22 crc kubenswrapper[4980]: I1206 03:35:22.536101 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:22 crc kubenswrapper[4980]: I1206 03:35:22.536117 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:22 crc kubenswrapper[4980]: I1206 03:35:22.536126 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:22Z","lastTransitionTime":"2025-12-06T03:35:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:22 crc kubenswrapper[4980]: I1206 03:35:22.638855 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:22 crc kubenswrapper[4980]: I1206 03:35:22.638888 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:22 crc kubenswrapper[4980]: I1206 03:35:22.638896 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:22 crc kubenswrapper[4980]: I1206 03:35:22.638909 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:22 crc kubenswrapper[4980]: I1206 03:35:22.638917 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:22Z","lastTransitionTime":"2025-12-06T03:35:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:22 crc kubenswrapper[4980]: I1206 03:35:22.742325 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:22 crc kubenswrapper[4980]: I1206 03:35:22.742403 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:22 crc kubenswrapper[4980]: I1206 03:35:22.742431 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:22 crc kubenswrapper[4980]: I1206 03:35:22.742463 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:22 crc kubenswrapper[4980]: I1206 03:35:22.742490 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:22Z","lastTransitionTime":"2025-12-06T03:35:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:22 crc kubenswrapper[4980]: I1206 03:35:22.845245 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:22 crc kubenswrapper[4980]: I1206 03:35:22.845277 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:22 crc kubenswrapper[4980]: I1206 03:35:22.845285 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:22 crc kubenswrapper[4980]: I1206 03:35:22.845297 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:22 crc kubenswrapper[4980]: I1206 03:35:22.845307 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:22Z","lastTransitionTime":"2025-12-06T03:35:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:22 crc kubenswrapper[4980]: I1206 03:35:22.948247 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:22 crc kubenswrapper[4980]: I1206 03:35:22.948323 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:22 crc kubenswrapper[4980]: I1206 03:35:22.948341 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:22 crc kubenswrapper[4980]: I1206 03:35:22.948368 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:22 crc kubenswrapper[4980]: I1206 03:35:22.948385 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:22Z","lastTransitionTime":"2025-12-06T03:35:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:23 crc kubenswrapper[4980]: I1206 03:35:23.052011 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:23 crc kubenswrapper[4980]: I1206 03:35:23.052090 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:23 crc kubenswrapper[4980]: I1206 03:35:23.052112 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:23 crc kubenswrapper[4980]: I1206 03:35:23.052148 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:23 crc kubenswrapper[4980]: I1206 03:35:23.052171 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:23Z","lastTransitionTime":"2025-12-06T03:35:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:23 crc kubenswrapper[4980]: I1206 03:35:23.155570 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:23 crc kubenswrapper[4980]: I1206 03:35:23.155623 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:23 crc kubenswrapper[4980]: I1206 03:35:23.155675 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:23 crc kubenswrapper[4980]: I1206 03:35:23.155698 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:23 crc kubenswrapper[4980]: I1206 03:35:23.155714 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:23Z","lastTransitionTime":"2025-12-06T03:35:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:23 crc kubenswrapper[4980]: I1206 03:35:23.162046 4980 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-5w4fr_5af0e768-3c9c-4401-ab56-57516cd3170e/ovnkube-controller/3.log" Dec 06 03:35:23 crc kubenswrapper[4980]: I1206 03:35:23.166287 4980 scope.go:117] "RemoveContainer" containerID="33fa9bd73b9af5df6de4e7ed019a512334f465b830a3f4396cde14989658bf55" Dec 06 03:35:23 crc kubenswrapper[4980]: E1206 03:35:23.166596 4980 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-5w4fr_openshift-ovn-kubernetes(5af0e768-3c9c-4401-ab56-57516cd3170e)\"" pod="openshift-ovn-kubernetes/ovnkube-node-5w4fr" podUID="5af0e768-3c9c-4401-ab56-57516cd3170e" Dec 06 03:35:23 crc kubenswrapper[4980]: I1206 03:35:23.184324 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 03:35:23 crc kubenswrapper[4980]: E1206 03:35:23.184482 4980 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 03:35:23 crc kubenswrapper[4980]: I1206 03:35:23.190968 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cab608b3-a02f-475b-9aae-9b8f5d878de9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f7fcc44b42bcbb55509048e3a566bb557333c4a83e33d7b0fba8eb514317909\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c22e7dd8886c2dfbb742900ed86ec7b0c2edf9a1363e74f6666961d928f06c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://473e792b2d0a9644854ee0581c6120fc9c989bc342899ef72b1d056080fd4694\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ee8e936dfc39361d6f99460910591ca8c901a20f0d274bbb6aaebfa8a00ee06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://77bc12768b018625ba11a6bfd66688f926c4b7f3bc1d0e41c6551a18b485ecf3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://27cf07cd96893900da4afecd0d56fd9269bc60c1d23246b4685fcd8756357c19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://27cf07cd96893900da4afecd0d56fd9269bc60c1d23246b4685fcd8756357c19\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c49feb24566a8284378670fa0390710d38c237acc94ffb654f90f3fa4626e4c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c49feb24566a8284378670fa0390710d38c237acc94ffb654f90f3fa4626e4c8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:03Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://bb74d1c3d752a379fa394c450217badb45511fbcafad62f7b618131cab94b29e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bb74d1c3d752a379fa394c450217badb45511fbcafad62f7b618131cab94b29e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:01Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:35:23Z is after 2025-08-24T17:21:41Z" Dec 06 03:35:23 crc kubenswrapper[4980]: I1206 03:35:23.204797 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://502c727d6baa31f400bee14c34b19abbe38af873f28f6a82a43547431755d981\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:35:23Z is after 2025-08-24T17:21:41Z" Dec 06 03:35:23 crc kubenswrapper[4980]: I1206 03:35:23.220360 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-dx44q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5bac22a-512c-49d0-90db-47de6aaaefd0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://97469dea2cd5393da0acf784ccc1969c8a4d07da7d51b14cdc955013ee9710ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f7mdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3b5dab476c7f227a6e4a2eae8469432667fb9fad345d9419ba59feab9fd385e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c3b5dab476c7f227a6e4a2eae8469432667fb9fad345d9419ba59feab9fd385e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f7mdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e921499231cbdf26b8918c98f9ad8c322fd7848865caeef048470de3e2c0efba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e921499231cbdf26b8918c98f9ad8c322fd7848865caeef048470de3e2c0efba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f7mdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7cbfe7d8cdb89916c238296592fab45b9e92e2c585baa6d67230decd1c84c5bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7cbfe7d8cdb89916c238296592fab45b9e92e2c585baa6d67230decd1c84c5bb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f7mdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f45cbf23b6b0d208cefc4b493f901bab9d71ccf2f1456507d54fb1f30c4fd48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7f45cbf23b6b0d208cefc4b493f901bab9d71ccf2f1456507d54fb1f30c4fd48\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f7mdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://344ce4b145ff021fcaf163803d8ad8cc6e9565bf7f0992e2653c43db663d8411\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://344ce4b145ff021fcaf163803d8ad8cc6e9565bf7f0992e2653c43db663d8411\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f7mdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://619e56a9937dd6376ab471ba16da30e5d77f443f5bd33610bea5ec85cd0446ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://619e56a9937dd6376ab471ba16da30e5d77f443f5bd33610bea5ec85cd0446ef\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f7mdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:23Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-dx44q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:35:23Z is after 2025-08-24T17:21:41Z" Dec 06 03:35:23 crc kubenswrapper[4980]: I1206 03:35:23.232779 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-r5zfb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"320f44d1-a671-4a91-b328-a8b0fdd8f23a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://92393ecaaef8cfcfd01837671f8cc89803dd570b637b4a373b40b7f487788826\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zhm5t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52e8db644463794041978ced2c0591c13e0ada8e5f1a78a94a0a665218a37b5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zhm5t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:23Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-r5zfb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:35:23Z is after 2025-08-24T17:21:41Z" Dec 06 03:35:23 crc kubenswrapper[4980]: I1206 03:35:23.245977 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-cvzg8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f1cde706-268d-4d26-818f-2f42bfb37b5f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://86f9a77acae360740be827eb0d1e1516e8a49cb9b197b6f5f9468ac70a101158\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rk86l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:29Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-cvzg8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:35:23Z is after 2025-08-24T17:21:41Z" Dec 06 03:35:23 crc kubenswrapper[4980]: I1206 03:35:23.258725 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:23 crc kubenswrapper[4980]: I1206 03:35:23.258761 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:23 crc kubenswrapper[4980]: I1206 03:35:23.258770 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:23 crc kubenswrapper[4980]: I1206 03:35:23.258784 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:23 crc kubenswrapper[4980]: I1206 03:35:23.258793 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:23Z","lastTransitionTime":"2025-12-06T03:35:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:23 crc kubenswrapper[4980]: I1206 03:35:23.259289 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:35:23Z is after 2025-08-24T17:21:41Z" Dec 06 03:35:23 crc kubenswrapper[4980]: I1206 03:35:23.275016 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4ef01e7d9809fd9cfc46e3a58b07159cdc6f5a3c210d1c9efc49bca27b30858\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:35:23Z is after 2025-08-24T17:21:41Z" Dec 06 03:35:23 crc kubenswrapper[4980]: I1206 03:35:23.287864 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-kq4z8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f96c1f6c-55e8-4d86-bff3-5381581b7022\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:35:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:35:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ea7f0bc23247b5effe7f0250582c8694bbf0e9a2d185a844773665a0a48f350e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://799498681afa7653408bc23b1e82d75c16e3c809d2ee7f322780e20c537eb710\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-06T03:35:15Z\\\",\\\"message\\\":\\\"2025-12-06T03:34:30+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_68ae7793-ff25-4e9c-8158-bea74df4ae16\\\\n2025-12-06T03:34:30+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_68ae7793-ff25-4e9c-8158-bea74df4ae16 to /host/opt/cni/bin/\\\\n2025-12-06T03:34:30Z [verbose] multus-daemon started\\\\n2025-12-06T03:34:30Z [verbose] Readiness Indicator file check\\\\n2025-12-06T03:35:15Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:24Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:35:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bdx7g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:23Z\\\"}}\" for pod \"openshift-multus\"/\"multus-kq4z8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:35:23Z is after 2025-08-24T17:21:41Z" Dec 06 03:35:23 crc kubenswrapper[4980]: I1206 03:35:23.298405 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-6kxfq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"951b6deb-bcea-4332-8ae8-c23ac7aaef36\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://423d30c3e65b04bdf6ff5ef53d8a7575a434c05e28489370a7370381ebc29de9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b8htf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://282df50dd9653c1cbe97fc2041d9e9ea91a992661682652aafc12e566d8f1b81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b8htf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:37Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-6kxfq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:35:23Z is after 2025-08-24T17:21:41Z" Dec 06 03:35:23 crc kubenswrapper[4980]: I1206 03:35:23.308460 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-mcprh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e6e88c5e-a255-449a-ae38-ac4d73a8e984\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g7xz9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g7xz9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:38Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-mcprh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:35:23Z is after 2025-08-24T17:21:41Z" Dec 06 03:35:23 crc kubenswrapper[4980]: I1206 03:35:23.321169 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7f958150-549d-4c80-8a7e-23c39e4f9500\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c28b6895b14e73f4f3eb7a8ce93076c05092ba9b58ea0e3e879c65bcfa227a21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://76cff8a2d652df38b04eefc053832baf4cbb5db778ff5e525b18c2825141b118\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://206f6f19fe3197f5d6798cc86c5473bf2decb02a2860ee7ffb5efb5f60952e8c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bafaa9670b05671fb57b658c2960bf3ee32c84e52f86a3a3bbab01d5a38ff0c4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://773a0bc2aab73ffc2717e987543defc9b786cc22300aa54b4ba4f3454504e331\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-06T03:34:22Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1206 03:34:15.305330 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1206 03:34:15.307249 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1592846546/tls.crt::/tmp/serving-cert-1592846546/tls.key\\\\\\\"\\\\nI1206 03:34:21.814037 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1206 03:34:21.816756 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1206 03:34:21.816780 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1206 03:34:21.816806 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1206 03:34:21.816813 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1206 03:34:21.831301 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1206 03:34:21.831311 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1206 03:34:21.831324 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 03:34:21.831330 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 03:34:21.831335 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1206 03:34:21.831339 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1206 03:34:21.831342 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1206 03:34:21.831346 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1206 03:34:21.911269 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:04Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d40e0e7446f8416bf9da94fc53c01be937f9a51a11eb101157d2ddd4597720d2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:04Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2cba89f2f1bed96d59f183628d900d9b72c91eb472e273c8bed0babcab485910\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2cba89f2f1bed96d59f183628d900d9b72c91eb472e273c8bed0babcab485910\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:01Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:35:23Z is after 2025-08-24T17:21:41Z" Dec 06 03:35:23 crc kubenswrapper[4980]: I1206 03:35:23.333417 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6ed68f66-1cf6-465a-8f66-76d5af6a44d8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://725e84f6bdfb722a8ce69158d03ea4c66a9f087642d55cb9651e003c81be0c8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://71f7b492ea07cdf7dc65fc48184ec1e64914eb1811bbfdddc0fb14c8ad6cb296\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b543aa4ee03d5a14909b67d876c5675c02e1df2a89a3c9ac8d3fb587fea53d8f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://32af3427e872a70cb3c2357116f04cabc8d3d9b69e7306b2e0de2ae22f9469d4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:01Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:35:23Z is after 2025-08-24T17:21:41Z" Dec 06 03:35:23 crc kubenswrapper[4980]: I1206 03:35:23.354455 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5w4fr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5af0e768-3c9c-4401-ab56-57516cd3170e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:23Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:23Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://768b8e1fe1a00a8467d744da5ad3f61006b0a3612c96f8e4d0ee6fcb93195f9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfzrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6342903b318317ee8d84c2f5d467cb25c811a789be0d6ab0e0cba37b5611e16d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfzrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d620685574db7594132ae24a91ec0a93dce20a9a289540deebb39650b8a1816\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfzrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c67ec8f9c1764fcdf215d52822c2d4f0437dc7f056f05d708767121f6c1f3515\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfzrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f35beae5339eeaf26494ed678e3c1bd02349eae2a4b4615325b17923f3b653b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfzrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a940ec6ee523313728f7e8ae934b7e0b0cff82771e3e64723309db1fae013c3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfzrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://33fa9bd73b9af5df6de4e7ed019a512334f465b830a3f4396cde14989658bf55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://33fa9bd73b9af5df6de4e7ed019a512334f465b830a3f4396cde14989658bf55\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-06T03:35:21Z\\\",\\\"message\\\":\\\"oller.go:360] Finished syncing service ingress-canary on namespace openshift-ingress-canary for network=default : 580.456µs\\\\nF1206 03:35:21.247527 6909 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:35:21Z is after 2025-08-24T17:21:41Z]\\\\nI1206 03:35:21.247533 6909 obj_retry.go:303] Retry object setup: *v1.Pod openshift-machine-config-operator/machine-config-daemon-r5zfb\\\\nI1206 03:35:21.247536 6909 obj_retry.go:303] Retry object setup: *v1.Pod openshift-network-operator/network-operator-58b4c7f79c-55gtf\\\\nI1206 03:35:21.247529 6909 obj_retry.go:303] Retry object setup: *v1.Pod openshift-network-console/networking-con\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T03:35:20Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-5w4fr_openshift-ovn-kubernetes(5af0e768-3c9c-4401-ab56-57516cd3170e)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfzrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe261f4322b85702181d7473ab2a9a4d2d4b869d586c3091f811b12a81a63aae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfzrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d876df04b5ab0ba07a5cbcafbe4648403c5014c59eb7d40adee2657c3a255473\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d876df04b5ab0ba07a5cbcafbe4648403c5014c59eb7d40adee2657c3a255473\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfzrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:23Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-5w4fr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:35:23Z is after 2025-08-24T17:21:41Z" Dec 06 03:35:23 crc kubenswrapper[4980]: I1206 03:35:23.361649 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:23 crc kubenswrapper[4980]: I1206 03:35:23.361695 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:23 crc kubenswrapper[4980]: I1206 03:35:23.361704 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:23 crc kubenswrapper[4980]: I1206 03:35:23.361721 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:23 crc kubenswrapper[4980]: I1206 03:35:23.361733 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:23Z","lastTransitionTime":"2025-12-06T03:35:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:23 crc kubenswrapper[4980]: I1206 03:35:23.372079 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"64278250-0b19-4591-b06e-6c96652ee92a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5beac3fd93b448506c0a110d70ef5faf42efa18b16abe6b3d4eb7b9c4233ad93\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ae66052702fcfd94f0123f4fbbe9efed37eb04609269a2bc951f95669de16601\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://173903f3554d7ac9158493498a016d618d92f46da1500408834b3c24249cf117\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1f596bfd5321ca225f4be03d4cf229c31a52def9731c0e177ab3a0bc010380e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1f596bfd5321ca225f4be03d4cf229c31a52def9731c0e177ab3a0bc010380e1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:02Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:01Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:35:23Z is after 2025-08-24T17:21:41Z" Dec 06 03:35:23 crc kubenswrapper[4980]: I1206 03:35:23.389923 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:22Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:35:23Z is after 2025-08-24T17:21:41Z" Dec 06 03:35:23 crc kubenswrapper[4980]: I1206 03:35:23.404374 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://334ea6a273fd7d268906c9f4a6dbf6f4df4fcdbf38f3d7d248cb7d6cf8537794\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ec44dcd5e4bf0a0466bb586ebd61259364dbbfe22499f4cbff854e3bd153082\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:35:23Z is after 2025-08-24T17:21:41Z" Dec 06 03:35:23 crc kubenswrapper[4980]: I1206 03:35:23.420856 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:22Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:35:23Z is after 2025-08-24T17:21:41Z" Dec 06 03:35:23 crc kubenswrapper[4980]: I1206 03:35:23.435231 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-sv5sz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"873977aa-4eb0-4526-bc0b-757ad2dea2a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01f0cc855c822a71ea8ef6a7ce6e25747ff9956691735973f935c4a28672a7a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lg7tx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:23Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-sv5sz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:35:23Z is after 2025-08-24T17:21:41Z" Dec 06 03:35:23 crc kubenswrapper[4980]: I1206 03:35:23.464205 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:23 crc kubenswrapper[4980]: I1206 03:35:23.464245 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:23 crc kubenswrapper[4980]: I1206 03:35:23.464256 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:23 crc kubenswrapper[4980]: I1206 03:35:23.464275 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:23 crc kubenswrapper[4980]: I1206 03:35:23.464287 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:23Z","lastTransitionTime":"2025-12-06T03:35:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:23 crc kubenswrapper[4980]: I1206 03:35:23.567404 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:23 crc kubenswrapper[4980]: I1206 03:35:23.567462 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:23 crc kubenswrapper[4980]: I1206 03:35:23.567472 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:23 crc kubenswrapper[4980]: I1206 03:35:23.567493 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:23 crc kubenswrapper[4980]: I1206 03:35:23.567508 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:23Z","lastTransitionTime":"2025-12-06T03:35:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:23 crc kubenswrapper[4980]: I1206 03:35:23.670282 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:23 crc kubenswrapper[4980]: I1206 03:35:23.670321 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:23 crc kubenswrapper[4980]: I1206 03:35:23.670333 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:23 crc kubenswrapper[4980]: I1206 03:35:23.670348 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:23 crc kubenswrapper[4980]: I1206 03:35:23.670361 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:23Z","lastTransitionTime":"2025-12-06T03:35:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:23 crc kubenswrapper[4980]: I1206 03:35:23.773503 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:23 crc kubenswrapper[4980]: I1206 03:35:23.773559 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:23 crc kubenswrapper[4980]: I1206 03:35:23.773567 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:23 crc kubenswrapper[4980]: I1206 03:35:23.773582 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:23 crc kubenswrapper[4980]: I1206 03:35:23.773591 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:23Z","lastTransitionTime":"2025-12-06T03:35:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:23 crc kubenswrapper[4980]: I1206 03:35:23.876649 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:23 crc kubenswrapper[4980]: I1206 03:35:23.876698 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:23 crc kubenswrapper[4980]: I1206 03:35:23.876709 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:23 crc kubenswrapper[4980]: I1206 03:35:23.876728 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:23 crc kubenswrapper[4980]: I1206 03:35:23.876740 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:23Z","lastTransitionTime":"2025-12-06T03:35:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:23 crc kubenswrapper[4980]: I1206 03:35:23.979505 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:23 crc kubenswrapper[4980]: I1206 03:35:23.979570 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:23 crc kubenswrapper[4980]: I1206 03:35:23.979583 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:23 crc kubenswrapper[4980]: I1206 03:35:23.979607 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:23 crc kubenswrapper[4980]: I1206 03:35:23.979619 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:23Z","lastTransitionTime":"2025-12-06T03:35:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:24 crc kubenswrapper[4980]: I1206 03:35:24.082729 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:24 crc kubenswrapper[4980]: I1206 03:35:24.082777 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:24 crc kubenswrapper[4980]: I1206 03:35:24.082787 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:24 crc kubenswrapper[4980]: I1206 03:35:24.082804 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:24 crc kubenswrapper[4980]: I1206 03:35:24.082812 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:24Z","lastTransitionTime":"2025-12-06T03:35:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:24 crc kubenswrapper[4980]: I1206 03:35:24.183812 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 03:35:24 crc kubenswrapper[4980]: I1206 03:35:24.183924 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-mcprh" Dec 06 03:35:24 crc kubenswrapper[4980]: I1206 03:35:24.184093 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 03:35:24 crc kubenswrapper[4980]: E1206 03:35:24.183939 4980 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 03:35:24 crc kubenswrapper[4980]: E1206 03:35:24.184486 4980 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-mcprh" podUID="e6e88c5e-a255-449a-ae38-ac4d73a8e984" Dec 06 03:35:24 crc kubenswrapper[4980]: E1206 03:35:24.184656 4980 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 03:35:24 crc kubenswrapper[4980]: I1206 03:35:24.185708 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:24 crc kubenswrapper[4980]: I1206 03:35:24.185742 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:24 crc kubenswrapper[4980]: I1206 03:35:24.185754 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:24 crc kubenswrapper[4980]: I1206 03:35:24.185771 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:24 crc kubenswrapper[4980]: I1206 03:35:24.185784 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:24Z","lastTransitionTime":"2025-12-06T03:35:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:24 crc kubenswrapper[4980]: I1206 03:35:24.288667 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:24 crc kubenswrapper[4980]: I1206 03:35:24.288708 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:24 crc kubenswrapper[4980]: I1206 03:35:24.288719 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:24 crc kubenswrapper[4980]: I1206 03:35:24.288736 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:24 crc kubenswrapper[4980]: I1206 03:35:24.288745 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:24Z","lastTransitionTime":"2025-12-06T03:35:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:24 crc kubenswrapper[4980]: I1206 03:35:24.391645 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:24 crc kubenswrapper[4980]: I1206 03:35:24.391712 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:24 crc kubenswrapper[4980]: I1206 03:35:24.391732 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:24 crc kubenswrapper[4980]: I1206 03:35:24.391757 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:24 crc kubenswrapper[4980]: I1206 03:35:24.391771 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:24Z","lastTransitionTime":"2025-12-06T03:35:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:24 crc kubenswrapper[4980]: I1206 03:35:24.494645 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:24 crc kubenswrapper[4980]: I1206 03:35:24.494705 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:24 crc kubenswrapper[4980]: I1206 03:35:24.494719 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:24 crc kubenswrapper[4980]: I1206 03:35:24.494744 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:24 crc kubenswrapper[4980]: I1206 03:35:24.494755 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:24Z","lastTransitionTime":"2025-12-06T03:35:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:24 crc kubenswrapper[4980]: I1206 03:35:24.597326 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:24 crc kubenswrapper[4980]: I1206 03:35:24.597393 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:24 crc kubenswrapper[4980]: I1206 03:35:24.597412 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:24 crc kubenswrapper[4980]: I1206 03:35:24.597439 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:24 crc kubenswrapper[4980]: I1206 03:35:24.597460 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:24Z","lastTransitionTime":"2025-12-06T03:35:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:24 crc kubenswrapper[4980]: I1206 03:35:24.701069 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:24 crc kubenswrapper[4980]: I1206 03:35:24.701137 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:24 crc kubenswrapper[4980]: I1206 03:35:24.701150 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:24 crc kubenswrapper[4980]: I1206 03:35:24.701174 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:24 crc kubenswrapper[4980]: I1206 03:35:24.701189 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:24Z","lastTransitionTime":"2025-12-06T03:35:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:24 crc kubenswrapper[4980]: I1206 03:35:24.804636 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:24 crc kubenswrapper[4980]: I1206 03:35:24.804703 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:24 crc kubenswrapper[4980]: I1206 03:35:24.804722 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:24 crc kubenswrapper[4980]: I1206 03:35:24.804749 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:24 crc kubenswrapper[4980]: I1206 03:35:24.804766 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:24Z","lastTransitionTime":"2025-12-06T03:35:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:24 crc kubenswrapper[4980]: I1206 03:35:24.908262 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:24 crc kubenswrapper[4980]: I1206 03:35:24.908305 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:24 crc kubenswrapper[4980]: I1206 03:35:24.908316 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:24 crc kubenswrapper[4980]: I1206 03:35:24.908334 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:24 crc kubenswrapper[4980]: I1206 03:35:24.908350 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:24Z","lastTransitionTime":"2025-12-06T03:35:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:25 crc kubenswrapper[4980]: I1206 03:35:25.011827 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:25 crc kubenswrapper[4980]: I1206 03:35:25.011910 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:25 crc kubenswrapper[4980]: I1206 03:35:25.011935 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:25 crc kubenswrapper[4980]: I1206 03:35:25.011966 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:25 crc kubenswrapper[4980]: I1206 03:35:25.011990 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:25Z","lastTransitionTime":"2025-12-06T03:35:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:25 crc kubenswrapper[4980]: I1206 03:35:25.115040 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:25 crc kubenswrapper[4980]: I1206 03:35:25.115770 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:25 crc kubenswrapper[4980]: I1206 03:35:25.115827 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:25 crc kubenswrapper[4980]: I1206 03:35:25.115861 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:25 crc kubenswrapper[4980]: I1206 03:35:25.115882 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:25Z","lastTransitionTime":"2025-12-06T03:35:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:25 crc kubenswrapper[4980]: I1206 03:35:25.184155 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 03:35:25 crc kubenswrapper[4980]: E1206 03:35:25.184379 4980 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 03:35:25 crc kubenswrapper[4980]: I1206 03:35:25.218448 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:25 crc kubenswrapper[4980]: I1206 03:35:25.218498 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:25 crc kubenswrapper[4980]: I1206 03:35:25.218526 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:25 crc kubenswrapper[4980]: I1206 03:35:25.218548 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:25 crc kubenswrapper[4980]: I1206 03:35:25.218563 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:25Z","lastTransitionTime":"2025-12-06T03:35:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:25 crc kubenswrapper[4980]: I1206 03:35:25.322351 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:25 crc kubenswrapper[4980]: I1206 03:35:25.322421 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:25 crc kubenswrapper[4980]: I1206 03:35:25.322438 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:25 crc kubenswrapper[4980]: I1206 03:35:25.322469 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:25 crc kubenswrapper[4980]: I1206 03:35:25.322487 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:25Z","lastTransitionTime":"2025-12-06T03:35:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:25 crc kubenswrapper[4980]: I1206 03:35:25.427054 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:25 crc kubenswrapper[4980]: I1206 03:35:25.431183 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:25 crc kubenswrapper[4980]: I1206 03:35:25.431223 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:25 crc kubenswrapper[4980]: I1206 03:35:25.431269 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:25 crc kubenswrapper[4980]: I1206 03:35:25.431297 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:25Z","lastTransitionTime":"2025-12-06T03:35:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:25 crc kubenswrapper[4980]: I1206 03:35:25.535161 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:25 crc kubenswrapper[4980]: I1206 03:35:25.535228 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:25 crc kubenswrapper[4980]: I1206 03:35:25.535252 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:25 crc kubenswrapper[4980]: I1206 03:35:25.535280 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:25 crc kubenswrapper[4980]: I1206 03:35:25.535300 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:25Z","lastTransitionTime":"2025-12-06T03:35:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:25 crc kubenswrapper[4980]: I1206 03:35:25.638550 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:25 crc kubenswrapper[4980]: I1206 03:35:25.638590 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:25 crc kubenswrapper[4980]: I1206 03:35:25.638600 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:25 crc kubenswrapper[4980]: I1206 03:35:25.638617 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:25 crc kubenswrapper[4980]: I1206 03:35:25.638630 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:25Z","lastTransitionTime":"2025-12-06T03:35:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:25 crc kubenswrapper[4980]: I1206 03:35:25.742216 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:25 crc kubenswrapper[4980]: I1206 03:35:25.742279 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:25 crc kubenswrapper[4980]: I1206 03:35:25.742299 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:25 crc kubenswrapper[4980]: I1206 03:35:25.742325 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:25 crc kubenswrapper[4980]: I1206 03:35:25.742349 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:25Z","lastTransitionTime":"2025-12-06T03:35:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:25 crc kubenswrapper[4980]: I1206 03:35:25.845782 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:25 crc kubenswrapper[4980]: I1206 03:35:25.845868 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:25 crc kubenswrapper[4980]: I1206 03:35:25.845893 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:25 crc kubenswrapper[4980]: I1206 03:35:25.845921 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:25 crc kubenswrapper[4980]: I1206 03:35:25.845942 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:25Z","lastTransitionTime":"2025-12-06T03:35:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:25 crc kubenswrapper[4980]: I1206 03:35:25.949689 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:25 crc kubenswrapper[4980]: I1206 03:35:25.949741 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:25 crc kubenswrapper[4980]: I1206 03:35:25.949750 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:25 crc kubenswrapper[4980]: I1206 03:35:25.949765 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:25 crc kubenswrapper[4980]: I1206 03:35:25.949775 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:25Z","lastTransitionTime":"2025-12-06T03:35:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:26 crc kubenswrapper[4980]: I1206 03:35:26.053433 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:26 crc kubenswrapper[4980]: I1206 03:35:26.053492 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:26 crc kubenswrapper[4980]: I1206 03:35:26.053530 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:26 crc kubenswrapper[4980]: I1206 03:35:26.053550 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:26 crc kubenswrapper[4980]: I1206 03:35:26.053562 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:26Z","lastTransitionTime":"2025-12-06T03:35:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:26 crc kubenswrapper[4980]: I1206 03:35:26.156983 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:26 crc kubenswrapper[4980]: I1206 03:35:26.157029 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:26 crc kubenswrapper[4980]: I1206 03:35:26.157042 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:26 crc kubenswrapper[4980]: I1206 03:35:26.157059 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:26 crc kubenswrapper[4980]: I1206 03:35:26.157069 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:26Z","lastTransitionTime":"2025-12-06T03:35:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:26 crc kubenswrapper[4980]: I1206 03:35:26.160626 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 03:35:26 crc kubenswrapper[4980]: I1206 03:35:26.160701 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 03:35:26 crc kubenswrapper[4980]: E1206 03:35:26.160833 4980 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 06 03:35:26 crc kubenswrapper[4980]: E1206 03:35:26.160943 4980 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 06 03:35:26 crc kubenswrapper[4980]: E1206 03:35:26.161248 4980 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-06 03:36:30.161202861 +0000 UTC m=+149.405954132 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 06 03:35:26 crc kubenswrapper[4980]: E1206 03:35:26.161310 4980 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-06 03:36:30.161278863 +0000 UTC m=+149.406030174 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 06 03:35:26 crc kubenswrapper[4980]: I1206 03:35:26.183949 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 03:35:26 crc kubenswrapper[4980]: I1206 03:35:26.184037 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 03:35:26 crc kubenswrapper[4980]: E1206 03:35:26.184198 4980 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 03:35:26 crc kubenswrapper[4980]: I1206 03:35:26.184245 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-mcprh" Dec 06 03:35:26 crc kubenswrapper[4980]: E1206 03:35:26.184324 4980 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-mcprh" podUID="e6e88c5e-a255-449a-ae38-ac4d73a8e984" Dec 06 03:35:26 crc kubenswrapper[4980]: E1206 03:35:26.184405 4980 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 03:35:26 crc kubenswrapper[4980]: I1206 03:35:26.260752 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:26 crc kubenswrapper[4980]: I1206 03:35:26.260823 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:26 crc kubenswrapper[4980]: I1206 03:35:26.260843 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:26 crc kubenswrapper[4980]: I1206 03:35:26.260871 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:26 crc kubenswrapper[4980]: I1206 03:35:26.260894 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:26Z","lastTransitionTime":"2025-12-06T03:35:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:26 crc kubenswrapper[4980]: I1206 03:35:26.361981 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 03:35:26 crc kubenswrapper[4980]: I1206 03:35:26.362129 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 03:35:26 crc kubenswrapper[4980]: I1206 03:35:26.362179 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 03:35:26 crc kubenswrapper[4980]: E1206 03:35:26.362348 4980 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 03:36:30.36231145 +0000 UTC m=+149.607062721 (durationBeforeRetry 1m4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 03:35:26 crc kubenswrapper[4980]: E1206 03:35:26.362419 4980 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 06 03:35:26 crc kubenswrapper[4980]: E1206 03:35:26.362466 4980 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 06 03:35:26 crc kubenswrapper[4980]: E1206 03:35:26.362486 4980 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 06 03:35:26 crc kubenswrapper[4980]: E1206 03:35:26.362576 4980 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 06 03:35:26 crc kubenswrapper[4980]: E1206 03:35:26.362617 4980 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-06 03:36:30.362592768 +0000 UTC m=+149.607344069 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 06 03:35:26 crc kubenswrapper[4980]: E1206 03:35:26.362628 4980 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 06 03:35:26 crc kubenswrapper[4980]: E1206 03:35:26.362657 4980 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 06 03:35:26 crc kubenswrapper[4980]: E1206 03:35:26.362746 4980 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-06 03:36:30.362719961 +0000 UTC m=+149.607471262 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 06 03:35:26 crc kubenswrapper[4980]: I1206 03:35:26.364041 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:26 crc kubenswrapper[4980]: I1206 03:35:26.364080 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:26 crc kubenswrapper[4980]: I1206 03:35:26.364095 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:26 crc kubenswrapper[4980]: I1206 03:35:26.364112 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:26 crc kubenswrapper[4980]: I1206 03:35:26.364127 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:26Z","lastTransitionTime":"2025-12-06T03:35:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:26 crc kubenswrapper[4980]: I1206 03:35:26.466104 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:26 crc kubenswrapper[4980]: I1206 03:35:26.466179 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:26 crc kubenswrapper[4980]: I1206 03:35:26.466208 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:26 crc kubenswrapper[4980]: I1206 03:35:26.466240 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:26 crc kubenswrapper[4980]: I1206 03:35:26.466262 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:26Z","lastTransitionTime":"2025-12-06T03:35:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:26 crc kubenswrapper[4980]: I1206 03:35:26.569433 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:26 crc kubenswrapper[4980]: I1206 03:35:26.569497 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:26 crc kubenswrapper[4980]: I1206 03:35:26.569541 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:26 crc kubenswrapper[4980]: I1206 03:35:26.569564 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:26 crc kubenswrapper[4980]: I1206 03:35:26.569591 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:26Z","lastTransitionTime":"2025-12-06T03:35:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:26 crc kubenswrapper[4980]: I1206 03:35:26.673222 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:26 crc kubenswrapper[4980]: I1206 03:35:26.673303 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:26 crc kubenswrapper[4980]: I1206 03:35:26.673327 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:26 crc kubenswrapper[4980]: I1206 03:35:26.673357 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:26 crc kubenswrapper[4980]: I1206 03:35:26.673383 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:26Z","lastTransitionTime":"2025-12-06T03:35:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:26 crc kubenswrapper[4980]: I1206 03:35:26.776445 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:26 crc kubenswrapper[4980]: I1206 03:35:26.776489 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:26 crc kubenswrapper[4980]: I1206 03:35:26.776508 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:26 crc kubenswrapper[4980]: I1206 03:35:26.776640 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:26 crc kubenswrapper[4980]: I1206 03:35:26.776659 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:26Z","lastTransitionTime":"2025-12-06T03:35:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:26 crc kubenswrapper[4980]: I1206 03:35:26.879672 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:26 crc kubenswrapper[4980]: I1206 03:35:26.879778 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:26 crc kubenswrapper[4980]: I1206 03:35:26.879802 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:26 crc kubenswrapper[4980]: I1206 03:35:26.879832 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:26 crc kubenswrapper[4980]: I1206 03:35:26.879859 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:26Z","lastTransitionTime":"2025-12-06T03:35:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:26 crc kubenswrapper[4980]: I1206 03:35:26.983929 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:26 crc kubenswrapper[4980]: I1206 03:35:26.983993 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:26 crc kubenswrapper[4980]: I1206 03:35:26.984013 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:26 crc kubenswrapper[4980]: I1206 03:35:26.984038 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:26 crc kubenswrapper[4980]: I1206 03:35:26.984059 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:26Z","lastTransitionTime":"2025-12-06T03:35:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:27 crc kubenswrapper[4980]: I1206 03:35:27.087690 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:27 crc kubenswrapper[4980]: I1206 03:35:27.088004 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:27 crc kubenswrapper[4980]: I1206 03:35:27.088018 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:27 crc kubenswrapper[4980]: I1206 03:35:27.088037 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:27 crc kubenswrapper[4980]: I1206 03:35:27.088049 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:27Z","lastTransitionTime":"2025-12-06T03:35:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:27 crc kubenswrapper[4980]: I1206 03:35:27.185205 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 03:35:27 crc kubenswrapper[4980]: E1206 03:35:27.185373 4980 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 03:35:27 crc kubenswrapper[4980]: I1206 03:35:27.191338 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:27 crc kubenswrapper[4980]: I1206 03:35:27.191420 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:27 crc kubenswrapper[4980]: I1206 03:35:27.191447 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:27 crc kubenswrapper[4980]: I1206 03:35:27.191482 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:27 crc kubenswrapper[4980]: I1206 03:35:27.191505 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:27Z","lastTransitionTime":"2025-12-06T03:35:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:27 crc kubenswrapper[4980]: I1206 03:35:27.199176 4980 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-crc"] Dec 06 03:35:27 crc kubenswrapper[4980]: I1206 03:35:27.295162 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:27 crc kubenswrapper[4980]: I1206 03:35:27.295227 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:27 crc kubenswrapper[4980]: I1206 03:35:27.295245 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:27 crc kubenswrapper[4980]: I1206 03:35:27.295270 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:27 crc kubenswrapper[4980]: I1206 03:35:27.295286 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:27Z","lastTransitionTime":"2025-12-06T03:35:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:27 crc kubenswrapper[4980]: I1206 03:35:27.398947 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:27 crc kubenswrapper[4980]: I1206 03:35:27.399015 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:27 crc kubenswrapper[4980]: I1206 03:35:27.399029 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:27 crc kubenswrapper[4980]: I1206 03:35:27.399059 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:27 crc kubenswrapper[4980]: I1206 03:35:27.399072 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:27Z","lastTransitionTime":"2025-12-06T03:35:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:27 crc kubenswrapper[4980]: I1206 03:35:27.502924 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:27 crc kubenswrapper[4980]: I1206 03:35:27.502986 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:27 crc kubenswrapper[4980]: I1206 03:35:27.502996 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:27 crc kubenswrapper[4980]: I1206 03:35:27.503015 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:27 crc kubenswrapper[4980]: I1206 03:35:27.503060 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:27Z","lastTransitionTime":"2025-12-06T03:35:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:27 crc kubenswrapper[4980]: I1206 03:35:27.607890 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:27 crc kubenswrapper[4980]: I1206 03:35:27.607971 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:27 crc kubenswrapper[4980]: I1206 03:35:27.607992 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:27 crc kubenswrapper[4980]: I1206 03:35:27.608021 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:27 crc kubenswrapper[4980]: I1206 03:35:27.608044 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:27Z","lastTransitionTime":"2025-12-06T03:35:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:27 crc kubenswrapper[4980]: I1206 03:35:27.711966 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:27 crc kubenswrapper[4980]: I1206 03:35:27.712055 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:27 crc kubenswrapper[4980]: I1206 03:35:27.712079 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:27 crc kubenswrapper[4980]: I1206 03:35:27.712111 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:27 crc kubenswrapper[4980]: I1206 03:35:27.712135 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:27Z","lastTransitionTime":"2025-12-06T03:35:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:27 crc kubenswrapper[4980]: I1206 03:35:27.815939 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:27 crc kubenswrapper[4980]: I1206 03:35:27.816018 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:27 crc kubenswrapper[4980]: I1206 03:35:27.816037 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:27 crc kubenswrapper[4980]: I1206 03:35:27.816066 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:27 crc kubenswrapper[4980]: I1206 03:35:27.816085 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:27Z","lastTransitionTime":"2025-12-06T03:35:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:27 crc kubenswrapper[4980]: I1206 03:35:27.919637 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:27 crc kubenswrapper[4980]: I1206 03:35:27.919714 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:27 crc kubenswrapper[4980]: I1206 03:35:27.919732 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:27 crc kubenswrapper[4980]: I1206 03:35:27.919759 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:27 crc kubenswrapper[4980]: I1206 03:35:27.919778 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:27Z","lastTransitionTime":"2025-12-06T03:35:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:28 crc kubenswrapper[4980]: I1206 03:35:28.023401 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:28 crc kubenswrapper[4980]: I1206 03:35:28.023484 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:28 crc kubenswrapper[4980]: I1206 03:35:28.023547 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:28 crc kubenswrapper[4980]: I1206 03:35:28.023584 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:28 crc kubenswrapper[4980]: I1206 03:35:28.023608 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:28Z","lastTransitionTime":"2025-12-06T03:35:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:28 crc kubenswrapper[4980]: I1206 03:35:28.127110 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:28 crc kubenswrapper[4980]: I1206 03:35:28.127223 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:28 crc kubenswrapper[4980]: I1206 03:35:28.127238 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:28 crc kubenswrapper[4980]: I1206 03:35:28.127255 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:28 crc kubenswrapper[4980]: I1206 03:35:28.127266 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:28Z","lastTransitionTime":"2025-12-06T03:35:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:28 crc kubenswrapper[4980]: I1206 03:35:28.183896 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 03:35:28 crc kubenswrapper[4980]: E1206 03:35:28.184043 4980 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 03:35:28 crc kubenswrapper[4980]: I1206 03:35:28.184203 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-mcprh" Dec 06 03:35:28 crc kubenswrapper[4980]: I1206 03:35:28.184268 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 03:35:28 crc kubenswrapper[4980]: E1206 03:35:28.184364 4980 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-mcprh" podUID="e6e88c5e-a255-449a-ae38-ac4d73a8e984" Dec 06 03:35:28 crc kubenswrapper[4980]: E1206 03:35:28.184582 4980 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 03:35:28 crc kubenswrapper[4980]: I1206 03:35:28.229731 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:28 crc kubenswrapper[4980]: I1206 03:35:28.229764 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:28 crc kubenswrapper[4980]: I1206 03:35:28.229775 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:28 crc kubenswrapper[4980]: I1206 03:35:28.229792 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:28 crc kubenswrapper[4980]: I1206 03:35:28.229804 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:28Z","lastTransitionTime":"2025-12-06T03:35:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:28 crc kubenswrapper[4980]: I1206 03:35:28.332815 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:28 crc kubenswrapper[4980]: I1206 03:35:28.332856 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:28 crc kubenswrapper[4980]: I1206 03:35:28.332864 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:28 crc kubenswrapper[4980]: I1206 03:35:28.332885 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:28 crc kubenswrapper[4980]: I1206 03:35:28.332896 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:28Z","lastTransitionTime":"2025-12-06T03:35:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:28 crc kubenswrapper[4980]: I1206 03:35:28.435620 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:28 crc kubenswrapper[4980]: I1206 03:35:28.435654 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:28 crc kubenswrapper[4980]: I1206 03:35:28.435663 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:28 crc kubenswrapper[4980]: I1206 03:35:28.435677 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:28 crc kubenswrapper[4980]: I1206 03:35:28.435686 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:28Z","lastTransitionTime":"2025-12-06T03:35:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:28 crc kubenswrapper[4980]: I1206 03:35:28.538177 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:28 crc kubenswrapper[4980]: I1206 03:35:28.538212 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:28 crc kubenswrapper[4980]: I1206 03:35:28.538221 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:28 crc kubenswrapper[4980]: I1206 03:35:28.538234 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:28 crc kubenswrapper[4980]: I1206 03:35:28.538243 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:28Z","lastTransitionTime":"2025-12-06T03:35:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:28 crc kubenswrapper[4980]: I1206 03:35:28.641187 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:28 crc kubenswrapper[4980]: I1206 03:35:28.641252 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:28 crc kubenswrapper[4980]: I1206 03:35:28.641271 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:28 crc kubenswrapper[4980]: I1206 03:35:28.641298 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:28 crc kubenswrapper[4980]: I1206 03:35:28.641317 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:28Z","lastTransitionTime":"2025-12-06T03:35:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:28 crc kubenswrapper[4980]: I1206 03:35:28.744192 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:28 crc kubenswrapper[4980]: I1206 03:35:28.744272 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:28 crc kubenswrapper[4980]: I1206 03:35:28.744292 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:28 crc kubenswrapper[4980]: I1206 03:35:28.744319 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:28 crc kubenswrapper[4980]: I1206 03:35:28.744337 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:28Z","lastTransitionTime":"2025-12-06T03:35:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:28 crc kubenswrapper[4980]: I1206 03:35:28.847054 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:28 crc kubenswrapper[4980]: I1206 03:35:28.847104 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:28 crc kubenswrapper[4980]: I1206 03:35:28.847118 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:28 crc kubenswrapper[4980]: I1206 03:35:28.847135 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:28 crc kubenswrapper[4980]: I1206 03:35:28.847145 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:28Z","lastTransitionTime":"2025-12-06T03:35:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:28 crc kubenswrapper[4980]: I1206 03:35:28.949926 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:28 crc kubenswrapper[4980]: I1206 03:35:28.950005 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:28 crc kubenswrapper[4980]: I1206 03:35:28.950024 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:28 crc kubenswrapper[4980]: I1206 03:35:28.950047 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:28 crc kubenswrapper[4980]: I1206 03:35:28.950065 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:28Z","lastTransitionTime":"2025-12-06T03:35:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:29 crc kubenswrapper[4980]: I1206 03:35:29.053336 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:29 crc kubenswrapper[4980]: I1206 03:35:29.053376 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:29 crc kubenswrapper[4980]: I1206 03:35:29.053388 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:29 crc kubenswrapper[4980]: I1206 03:35:29.053404 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:29 crc kubenswrapper[4980]: I1206 03:35:29.053415 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:29Z","lastTransitionTime":"2025-12-06T03:35:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:29 crc kubenswrapper[4980]: I1206 03:35:29.156272 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:29 crc kubenswrapper[4980]: I1206 03:35:29.156313 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:29 crc kubenswrapper[4980]: I1206 03:35:29.156323 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:29 crc kubenswrapper[4980]: I1206 03:35:29.156345 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:29 crc kubenswrapper[4980]: I1206 03:35:29.156377 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:29Z","lastTransitionTime":"2025-12-06T03:35:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:29 crc kubenswrapper[4980]: I1206 03:35:29.183929 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 03:35:29 crc kubenswrapper[4980]: E1206 03:35:29.184125 4980 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 03:35:29 crc kubenswrapper[4980]: I1206 03:35:29.258779 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:29 crc kubenswrapper[4980]: I1206 03:35:29.258825 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:29 crc kubenswrapper[4980]: I1206 03:35:29.258838 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:29 crc kubenswrapper[4980]: I1206 03:35:29.258868 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:29 crc kubenswrapper[4980]: I1206 03:35:29.258889 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:29Z","lastTransitionTime":"2025-12-06T03:35:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:29 crc kubenswrapper[4980]: I1206 03:35:29.362058 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:29 crc kubenswrapper[4980]: I1206 03:35:29.362135 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:29 crc kubenswrapper[4980]: I1206 03:35:29.362154 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:29 crc kubenswrapper[4980]: I1206 03:35:29.362179 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:29 crc kubenswrapper[4980]: I1206 03:35:29.362196 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:29Z","lastTransitionTime":"2025-12-06T03:35:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:29 crc kubenswrapper[4980]: I1206 03:35:29.468630 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:29 crc kubenswrapper[4980]: I1206 03:35:29.468711 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:29 crc kubenswrapper[4980]: I1206 03:35:29.468736 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:29 crc kubenswrapper[4980]: I1206 03:35:29.468769 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:29 crc kubenswrapper[4980]: I1206 03:35:29.468793 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:29Z","lastTransitionTime":"2025-12-06T03:35:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:29 crc kubenswrapper[4980]: I1206 03:35:29.573115 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:29 crc kubenswrapper[4980]: I1206 03:35:29.573178 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:29 crc kubenswrapper[4980]: I1206 03:35:29.573195 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:29 crc kubenswrapper[4980]: I1206 03:35:29.573219 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:29 crc kubenswrapper[4980]: I1206 03:35:29.573241 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:29Z","lastTransitionTime":"2025-12-06T03:35:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:29 crc kubenswrapper[4980]: I1206 03:35:29.676892 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:29 crc kubenswrapper[4980]: I1206 03:35:29.676959 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:29 crc kubenswrapper[4980]: I1206 03:35:29.676984 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:29 crc kubenswrapper[4980]: I1206 03:35:29.677015 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:29 crc kubenswrapper[4980]: I1206 03:35:29.677042 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:29Z","lastTransitionTime":"2025-12-06T03:35:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:29 crc kubenswrapper[4980]: I1206 03:35:29.780779 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:29 crc kubenswrapper[4980]: I1206 03:35:29.780857 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:29 crc kubenswrapper[4980]: I1206 03:35:29.780885 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:29 crc kubenswrapper[4980]: I1206 03:35:29.780917 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:29 crc kubenswrapper[4980]: I1206 03:35:29.780941 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:29Z","lastTransitionTime":"2025-12-06T03:35:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:29 crc kubenswrapper[4980]: I1206 03:35:29.884824 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:29 crc kubenswrapper[4980]: I1206 03:35:29.884945 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:29 crc kubenswrapper[4980]: I1206 03:35:29.884972 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:29 crc kubenswrapper[4980]: I1206 03:35:29.885003 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:29 crc kubenswrapper[4980]: I1206 03:35:29.885028 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:29Z","lastTransitionTime":"2025-12-06T03:35:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:29 crc kubenswrapper[4980]: I1206 03:35:29.988235 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:29 crc kubenswrapper[4980]: I1206 03:35:29.988298 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:29 crc kubenswrapper[4980]: I1206 03:35:29.988315 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:29 crc kubenswrapper[4980]: I1206 03:35:29.988341 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:29 crc kubenswrapper[4980]: I1206 03:35:29.988362 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:29Z","lastTransitionTime":"2025-12-06T03:35:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:30 crc kubenswrapper[4980]: I1206 03:35:30.091777 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:30 crc kubenswrapper[4980]: I1206 03:35:30.091838 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:30 crc kubenswrapper[4980]: I1206 03:35:30.091856 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:30 crc kubenswrapper[4980]: I1206 03:35:30.091879 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:30 crc kubenswrapper[4980]: I1206 03:35:30.091895 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:30Z","lastTransitionTime":"2025-12-06T03:35:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:30 crc kubenswrapper[4980]: I1206 03:35:30.120480 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:30 crc kubenswrapper[4980]: I1206 03:35:30.120579 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:30 crc kubenswrapper[4980]: I1206 03:35:30.120611 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:30 crc kubenswrapper[4980]: I1206 03:35:30.120641 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:30 crc kubenswrapper[4980]: I1206 03:35:30.120660 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:30Z","lastTransitionTime":"2025-12-06T03:35:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:30 crc kubenswrapper[4980]: E1206 03:35:30.144030 4980 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T03:35:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T03:35:30Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T03:35:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T03:35:30Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T03:35:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T03:35:30Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T03:35:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T03:35:30Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"9b4ed4f8-ed8d-447b-9e1f-13defbdcf0f7\\\",\\\"systemUUID\\\":\\\"a5b493aa-69de-4688-8ff6-1e5f49b7f014\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:35:30Z is after 2025-08-24T17:21:41Z" Dec 06 03:35:30 crc kubenswrapper[4980]: I1206 03:35:30.150055 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:30 crc kubenswrapper[4980]: I1206 03:35:30.150136 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:30 crc kubenswrapper[4980]: I1206 03:35:30.150186 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:30 crc kubenswrapper[4980]: I1206 03:35:30.150222 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:30 crc kubenswrapper[4980]: I1206 03:35:30.150246 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:30Z","lastTransitionTime":"2025-12-06T03:35:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:30 crc kubenswrapper[4980]: E1206 03:35:30.176892 4980 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T03:35:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T03:35:30Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T03:35:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T03:35:30Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T03:35:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T03:35:30Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T03:35:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T03:35:30Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"9b4ed4f8-ed8d-447b-9e1f-13defbdcf0f7\\\",\\\"systemUUID\\\":\\\"a5b493aa-69de-4688-8ff6-1e5f49b7f014\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:35:30Z is after 2025-08-24T17:21:41Z" Dec 06 03:35:30 crc kubenswrapper[4980]: I1206 03:35:30.182746 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:30 crc kubenswrapper[4980]: I1206 03:35:30.182815 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:30 crc kubenswrapper[4980]: I1206 03:35:30.182832 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:30 crc kubenswrapper[4980]: I1206 03:35:30.182857 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:30 crc kubenswrapper[4980]: I1206 03:35:30.182877 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:30Z","lastTransitionTime":"2025-12-06T03:35:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:30 crc kubenswrapper[4980]: I1206 03:35:30.183983 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-mcprh" Dec 06 03:35:30 crc kubenswrapper[4980]: I1206 03:35:30.183978 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 03:35:30 crc kubenswrapper[4980]: E1206 03:35:30.184443 4980 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 03:35:30 crc kubenswrapper[4980]: I1206 03:35:30.184063 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 03:35:30 crc kubenswrapper[4980]: E1206 03:35:30.184443 4980 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-mcprh" podUID="e6e88c5e-a255-449a-ae38-ac4d73a8e984" Dec 06 03:35:30 crc kubenswrapper[4980]: E1206 03:35:30.184593 4980 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 03:35:30 crc kubenswrapper[4980]: E1206 03:35:30.201033 4980 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T03:35:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T03:35:30Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T03:35:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T03:35:30Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T03:35:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T03:35:30Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T03:35:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T03:35:30Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"9b4ed4f8-ed8d-447b-9e1f-13defbdcf0f7\\\",\\\"systemUUID\\\":\\\"a5b493aa-69de-4688-8ff6-1e5f49b7f014\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:35:30Z is after 2025-08-24T17:21:41Z" Dec 06 03:35:30 crc kubenswrapper[4980]: I1206 03:35:30.206839 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:30 crc kubenswrapper[4980]: I1206 03:35:30.207007 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:30 crc kubenswrapper[4980]: I1206 03:35:30.207121 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:30 crc kubenswrapper[4980]: I1206 03:35:30.207221 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:30 crc kubenswrapper[4980]: I1206 03:35:30.207308 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:30Z","lastTransitionTime":"2025-12-06T03:35:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:30 crc kubenswrapper[4980]: E1206 03:35:30.224128 4980 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T03:35:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T03:35:30Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T03:35:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T03:35:30Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T03:35:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T03:35:30Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T03:35:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T03:35:30Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"9b4ed4f8-ed8d-447b-9e1f-13defbdcf0f7\\\",\\\"systemUUID\\\":\\\"a5b493aa-69de-4688-8ff6-1e5f49b7f014\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:35:30Z is after 2025-08-24T17:21:41Z" Dec 06 03:35:30 crc kubenswrapper[4980]: I1206 03:35:30.230066 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:30 crc kubenswrapper[4980]: I1206 03:35:30.230291 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:30 crc kubenswrapper[4980]: I1206 03:35:30.230317 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:30 crc kubenswrapper[4980]: I1206 03:35:30.230341 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:30 crc kubenswrapper[4980]: I1206 03:35:30.230354 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:30Z","lastTransitionTime":"2025-12-06T03:35:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:30 crc kubenswrapper[4980]: E1206 03:35:30.248308 4980 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T03:35:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T03:35:30Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T03:35:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T03:35:30Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T03:35:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T03:35:30Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T03:35:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T03:35:30Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"9b4ed4f8-ed8d-447b-9e1f-13defbdcf0f7\\\",\\\"systemUUID\\\":\\\"a5b493aa-69de-4688-8ff6-1e5f49b7f014\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:35:30Z is after 2025-08-24T17:21:41Z" Dec 06 03:35:30 crc kubenswrapper[4980]: E1206 03:35:30.248678 4980 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 06 03:35:30 crc kubenswrapper[4980]: I1206 03:35:30.251012 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:30 crc kubenswrapper[4980]: I1206 03:35:30.251061 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:30 crc kubenswrapper[4980]: I1206 03:35:30.251074 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:30 crc kubenswrapper[4980]: I1206 03:35:30.251097 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:30 crc kubenswrapper[4980]: I1206 03:35:30.251109 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:30Z","lastTransitionTime":"2025-12-06T03:35:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:30 crc kubenswrapper[4980]: I1206 03:35:30.353728 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:30 crc kubenswrapper[4980]: I1206 03:35:30.353837 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:30 crc kubenswrapper[4980]: I1206 03:35:30.353859 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:30 crc kubenswrapper[4980]: I1206 03:35:30.353885 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:30 crc kubenswrapper[4980]: I1206 03:35:30.353904 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:30Z","lastTransitionTime":"2025-12-06T03:35:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:30 crc kubenswrapper[4980]: I1206 03:35:30.457026 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:30 crc kubenswrapper[4980]: I1206 03:35:30.457170 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:30 crc kubenswrapper[4980]: I1206 03:35:30.457191 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:30 crc kubenswrapper[4980]: I1206 03:35:30.457218 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:30 crc kubenswrapper[4980]: I1206 03:35:30.457240 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:30Z","lastTransitionTime":"2025-12-06T03:35:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:30 crc kubenswrapper[4980]: I1206 03:35:30.560384 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:30 crc kubenswrapper[4980]: I1206 03:35:30.560466 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:30 crc kubenswrapper[4980]: I1206 03:35:30.560482 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:30 crc kubenswrapper[4980]: I1206 03:35:30.560508 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:30 crc kubenswrapper[4980]: I1206 03:35:30.560552 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:30Z","lastTransitionTime":"2025-12-06T03:35:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:30 crc kubenswrapper[4980]: I1206 03:35:30.663643 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:30 crc kubenswrapper[4980]: I1206 03:35:30.663717 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:30 crc kubenswrapper[4980]: I1206 03:35:30.663743 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:30 crc kubenswrapper[4980]: I1206 03:35:30.663773 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:30 crc kubenswrapper[4980]: I1206 03:35:30.663796 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:30Z","lastTransitionTime":"2025-12-06T03:35:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:30 crc kubenswrapper[4980]: I1206 03:35:30.767458 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:30 crc kubenswrapper[4980]: I1206 03:35:30.767813 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:30 crc kubenswrapper[4980]: I1206 03:35:30.767974 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:30 crc kubenswrapper[4980]: I1206 03:35:30.768168 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:30 crc kubenswrapper[4980]: I1206 03:35:30.768324 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:30Z","lastTransitionTime":"2025-12-06T03:35:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:30 crc kubenswrapper[4980]: I1206 03:35:30.872332 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:30 crc kubenswrapper[4980]: I1206 03:35:30.872710 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:30 crc kubenswrapper[4980]: I1206 03:35:30.872965 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:30 crc kubenswrapper[4980]: I1206 03:35:30.873173 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:30 crc kubenswrapper[4980]: I1206 03:35:30.873335 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:30Z","lastTransitionTime":"2025-12-06T03:35:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:30 crc kubenswrapper[4980]: I1206 03:35:30.976383 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:30 crc kubenswrapper[4980]: I1206 03:35:30.976435 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:30 crc kubenswrapper[4980]: I1206 03:35:30.976473 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:30 crc kubenswrapper[4980]: I1206 03:35:30.976558 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:30 crc kubenswrapper[4980]: I1206 03:35:30.976584 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:30Z","lastTransitionTime":"2025-12-06T03:35:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:31 crc kubenswrapper[4980]: I1206 03:35:31.080301 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:31 crc kubenswrapper[4980]: I1206 03:35:31.080676 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:31 crc kubenswrapper[4980]: I1206 03:35:31.080918 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:31 crc kubenswrapper[4980]: I1206 03:35:31.081147 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:31 crc kubenswrapper[4980]: I1206 03:35:31.081349 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:31Z","lastTransitionTime":"2025-12-06T03:35:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:31 crc kubenswrapper[4980]: I1206 03:35:31.183416 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 03:35:31 crc kubenswrapper[4980]: E1206 03:35:31.184088 4980 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 03:35:31 crc kubenswrapper[4980]: I1206 03:35:31.185551 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:31 crc kubenswrapper[4980]: I1206 03:35:31.185593 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:31 crc kubenswrapper[4980]: I1206 03:35:31.185602 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:31 crc kubenswrapper[4980]: I1206 03:35:31.185617 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:31 crc kubenswrapper[4980]: I1206 03:35:31.185626 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:31Z","lastTransitionTime":"2025-12-06T03:35:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:31 crc kubenswrapper[4980]: I1206 03:35:31.204894 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"64278250-0b19-4591-b06e-6c96652ee92a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5beac3fd93b448506c0a110d70ef5faf42efa18b16abe6b3d4eb7b9c4233ad93\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ae66052702fcfd94f0123f4fbbe9efed37eb04609269a2bc951f95669de16601\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://173903f3554d7ac9158493498a016d618d92f46da1500408834b3c24249cf117\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1f596bfd5321ca225f4be03d4cf229c31a52def9731c0e177ab3a0bc010380e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1f596bfd5321ca225f4be03d4cf229c31a52def9731c0e177ab3a0bc010380e1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T03:34:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T03:34:02Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T03:34:01Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:35:31Z is after 2025-08-24T17:21:41Z" Dec 06 03:35:31 crc kubenswrapper[4980]: I1206 03:35:31.224055 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:22Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:35:31Z is after 2025-08-24T17:21:41Z" Dec 06 03:35:31 crc kubenswrapper[4980]: I1206 03:35:31.243726 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://334ea6a273fd7d268906c9f4a6dbf6f4df4fcdbf38f3d7d248cb7d6cf8537794\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ec44dcd5e4bf0a0466bb586ebd61259364dbbfe22499f4cbff854e3bd153082\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T03:34:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:35:31Z is after 2025-08-24T17:21:41Z" Dec 06 03:35:31 crc kubenswrapper[4980]: I1206 03:35:31.255846 4980 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T03:34:22Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T03:35:31Z is after 2025-08-24T17:21:41Z" Dec 06 03:35:31 crc kubenswrapper[4980]: I1206 03:35:31.276275 4980 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/node-resolver-sv5sz" podStartSLOduration=70.276238713 podStartE2EDuration="1m10.276238713s" podCreationTimestamp="2025-12-06 03:34:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 03:35:31.276223183 +0000 UTC m=+90.520974504" watchObservedRunningTime="2025-12-06 03:35:31.276238713 +0000 UTC m=+90.520989984" Dec 06 03:35:31 crc kubenswrapper[4980]: I1206 03:35:31.287693 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:31 crc kubenswrapper[4980]: I1206 03:35:31.287730 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:31 crc kubenswrapper[4980]: I1206 03:35:31.287744 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:31 crc kubenswrapper[4980]: I1206 03:35:31.287764 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:31 crc kubenswrapper[4980]: I1206 03:35:31.287777 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:31Z","lastTransitionTime":"2025-12-06T03:35:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:31 crc kubenswrapper[4980]: I1206 03:35:31.351275 4980 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd/etcd-crc" podStartSLOduration=70.35124256 podStartE2EDuration="1m10.35124256s" podCreationTimestamp="2025-12-06 03:34:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 03:35:31.332040768 +0000 UTC m=+90.576792059" watchObservedRunningTime="2025-12-06 03:35:31.35124256 +0000 UTC m=+90.595993861" Dec 06 03:35:31 crc kubenswrapper[4980]: I1206 03:35:31.382632 4980 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-additional-cni-plugins-dx44q" podStartSLOduration=69.382602978 podStartE2EDuration="1m9.382602978s" podCreationTimestamp="2025-12-06 03:34:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 03:35:31.382240398 +0000 UTC m=+90.626991709" watchObservedRunningTime="2025-12-06 03:35:31.382602978 +0000 UTC m=+90.627354249" Dec 06 03:35:31 crc kubenswrapper[4980]: I1206 03:35:31.392030 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:31 crc kubenswrapper[4980]: I1206 03:35:31.392077 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:31 crc kubenswrapper[4980]: I1206 03:35:31.392087 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:31 crc kubenswrapper[4980]: I1206 03:35:31.392104 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:31 crc kubenswrapper[4980]: I1206 03:35:31.392115 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:31Z","lastTransitionTime":"2025-12-06T03:35:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:31 crc kubenswrapper[4980]: I1206 03:35:31.415611 4980 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-daemon-r5zfb" podStartSLOduration=69.415587832 podStartE2EDuration="1m9.415587832s" podCreationTimestamp="2025-12-06 03:34:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 03:35:31.399919798 +0000 UTC m=+90.644671109" watchObservedRunningTime="2025-12-06 03:35:31.415587832 +0000 UTC m=+90.660339103" Dec 06 03:35:31 crc kubenswrapper[4980]: I1206 03:35:31.415724 4980 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/node-ca-cvzg8" podStartSLOduration=69.415718555 podStartE2EDuration="1m9.415718555s" podCreationTimestamp="2025-12-06 03:34:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 03:35:31.415585672 +0000 UTC m=+90.660336943" watchObservedRunningTime="2025-12-06 03:35:31.415718555 +0000 UTC m=+90.660469826" Dec 06 03:35:31 crc kubenswrapper[4980]: I1206 03:35:31.494766 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:31 crc kubenswrapper[4980]: I1206 03:35:31.494807 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:31 crc kubenswrapper[4980]: I1206 03:35:31.494816 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:31 crc kubenswrapper[4980]: I1206 03:35:31.494834 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:31 crc kubenswrapper[4980]: I1206 03:35:31.494845 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:31Z","lastTransitionTime":"2025-12-06T03:35:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:31 crc kubenswrapper[4980]: I1206 03:35:31.505266 4980 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-6kxfq" podStartSLOduration=68.505251415 podStartE2EDuration="1m8.505251415s" podCreationTimestamp="2025-12-06 03:34:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 03:35:31.504628297 +0000 UTC m=+90.749379568" watchObservedRunningTime="2025-12-06 03:35:31.505251415 +0000 UTC m=+90.750002686" Dec 06 03:35:31 crc kubenswrapper[4980]: I1206 03:35:31.505434 4980 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-kq4z8" podStartSLOduration=69.50543011 podStartE2EDuration="1m9.50543011s" podCreationTimestamp="2025-12-06 03:34:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 03:35:31.486758483 +0000 UTC m=+90.731509754" watchObservedRunningTime="2025-12-06 03:35:31.50543011 +0000 UTC m=+90.750181371" Dec 06 03:35:31 crc kubenswrapper[4980]: I1206 03:35:31.543102 4980 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" podStartSLOduration=4.543074312 podStartE2EDuration="4.543074312s" podCreationTimestamp="2025-12-06 03:35:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 03:35:31.542101145 +0000 UTC m=+90.786852416" watchObservedRunningTime="2025-12-06 03:35:31.543074312 +0000 UTC m=+90.787825583" Dec 06 03:35:31 crc kubenswrapper[4980]: I1206 03:35:31.575364 4980 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podStartSLOduration=63.575344816 podStartE2EDuration="1m3.575344816s" podCreationTimestamp="2025-12-06 03:34:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 03:35:31.574543913 +0000 UTC m=+90.819295194" watchObservedRunningTime="2025-12-06 03:35:31.575344816 +0000 UTC m=+90.820096087" Dec 06 03:35:31 crc kubenswrapper[4980]: I1206 03:35:31.575475 4980 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=69.575471049 podStartE2EDuration="1m9.575471049s" podCreationTimestamp="2025-12-06 03:34:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 03:35:31.55854569 +0000 UTC m=+90.803296971" watchObservedRunningTime="2025-12-06 03:35:31.575471049 +0000 UTC m=+90.820222320" Dec 06 03:35:31 crc kubenswrapper[4980]: I1206 03:35:31.597900 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:31 crc kubenswrapper[4980]: I1206 03:35:31.597966 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:31 crc kubenswrapper[4980]: I1206 03:35:31.597981 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:31 crc kubenswrapper[4980]: I1206 03:35:31.598006 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:31 crc kubenswrapper[4980]: I1206 03:35:31.598019 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:31Z","lastTransitionTime":"2025-12-06T03:35:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:31 crc kubenswrapper[4980]: I1206 03:35:31.699890 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:31 crc kubenswrapper[4980]: I1206 03:35:31.699937 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:31 crc kubenswrapper[4980]: I1206 03:35:31.699951 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:31 crc kubenswrapper[4980]: I1206 03:35:31.699969 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:31 crc kubenswrapper[4980]: I1206 03:35:31.699983 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:31Z","lastTransitionTime":"2025-12-06T03:35:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:31 crc kubenswrapper[4980]: I1206 03:35:31.802573 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:31 crc kubenswrapper[4980]: I1206 03:35:31.802603 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:31 crc kubenswrapper[4980]: I1206 03:35:31.802613 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:31 crc kubenswrapper[4980]: I1206 03:35:31.802627 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:31 crc kubenswrapper[4980]: I1206 03:35:31.802638 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:31Z","lastTransitionTime":"2025-12-06T03:35:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:31 crc kubenswrapper[4980]: I1206 03:35:31.904714 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:31 crc kubenswrapper[4980]: I1206 03:35:31.904772 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:31 crc kubenswrapper[4980]: I1206 03:35:31.904823 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:31 crc kubenswrapper[4980]: I1206 03:35:31.904846 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:31 crc kubenswrapper[4980]: I1206 03:35:31.904863 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:31Z","lastTransitionTime":"2025-12-06T03:35:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:32 crc kubenswrapper[4980]: I1206 03:35:32.008217 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:32 crc kubenswrapper[4980]: I1206 03:35:32.008264 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:32 crc kubenswrapper[4980]: I1206 03:35:32.008273 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:32 crc kubenswrapper[4980]: I1206 03:35:32.008288 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:32 crc kubenswrapper[4980]: I1206 03:35:32.008299 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:32Z","lastTransitionTime":"2025-12-06T03:35:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:32 crc kubenswrapper[4980]: I1206 03:35:32.111991 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:32 crc kubenswrapper[4980]: I1206 03:35:32.112051 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:32 crc kubenswrapper[4980]: I1206 03:35:32.112064 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:32 crc kubenswrapper[4980]: I1206 03:35:32.112087 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:32 crc kubenswrapper[4980]: I1206 03:35:32.112108 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:32Z","lastTransitionTime":"2025-12-06T03:35:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:32 crc kubenswrapper[4980]: I1206 03:35:32.183463 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-mcprh" Dec 06 03:35:32 crc kubenswrapper[4980]: E1206 03:35:32.183650 4980 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-mcprh" podUID="e6e88c5e-a255-449a-ae38-ac4d73a8e984" Dec 06 03:35:32 crc kubenswrapper[4980]: I1206 03:35:32.183649 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 03:35:32 crc kubenswrapper[4980]: I1206 03:35:32.183689 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 03:35:32 crc kubenswrapper[4980]: E1206 03:35:32.183737 4980 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 03:35:32 crc kubenswrapper[4980]: E1206 03:35:32.183906 4980 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 03:35:32 crc kubenswrapper[4980]: I1206 03:35:32.214107 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:32 crc kubenswrapper[4980]: I1206 03:35:32.214155 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:32 crc kubenswrapper[4980]: I1206 03:35:32.214167 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:32 crc kubenswrapper[4980]: I1206 03:35:32.214183 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:32 crc kubenswrapper[4980]: I1206 03:35:32.214195 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:32Z","lastTransitionTime":"2025-12-06T03:35:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:32 crc kubenswrapper[4980]: I1206 03:35:32.318496 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:32 crc kubenswrapper[4980]: I1206 03:35:32.318971 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:32 crc kubenswrapper[4980]: I1206 03:35:32.319129 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:32 crc kubenswrapper[4980]: I1206 03:35:32.319285 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:32 crc kubenswrapper[4980]: I1206 03:35:32.319445 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:32Z","lastTransitionTime":"2025-12-06T03:35:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:32 crc kubenswrapper[4980]: I1206 03:35:32.422087 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:32 crc kubenswrapper[4980]: I1206 03:35:32.422134 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:32 crc kubenswrapper[4980]: I1206 03:35:32.422145 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:32 crc kubenswrapper[4980]: I1206 03:35:32.422162 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:32 crc kubenswrapper[4980]: I1206 03:35:32.422177 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:32Z","lastTransitionTime":"2025-12-06T03:35:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:32 crc kubenswrapper[4980]: I1206 03:35:32.525050 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:32 crc kubenswrapper[4980]: I1206 03:35:32.525121 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:32 crc kubenswrapper[4980]: I1206 03:35:32.525143 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:32 crc kubenswrapper[4980]: I1206 03:35:32.525170 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:32 crc kubenswrapper[4980]: I1206 03:35:32.525191 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:32Z","lastTransitionTime":"2025-12-06T03:35:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:32 crc kubenswrapper[4980]: I1206 03:35:32.628714 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:32 crc kubenswrapper[4980]: I1206 03:35:32.628778 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:32 crc kubenswrapper[4980]: I1206 03:35:32.628798 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:32 crc kubenswrapper[4980]: I1206 03:35:32.628822 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:32 crc kubenswrapper[4980]: I1206 03:35:32.628841 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:32Z","lastTransitionTime":"2025-12-06T03:35:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:32 crc kubenswrapper[4980]: I1206 03:35:32.732156 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:32 crc kubenswrapper[4980]: I1206 03:35:32.732229 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:32 crc kubenswrapper[4980]: I1206 03:35:32.732252 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:32 crc kubenswrapper[4980]: I1206 03:35:32.732281 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:32 crc kubenswrapper[4980]: I1206 03:35:32.732301 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:32Z","lastTransitionTime":"2025-12-06T03:35:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:32 crc kubenswrapper[4980]: I1206 03:35:32.834678 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:32 crc kubenswrapper[4980]: I1206 03:35:32.834783 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:32 crc kubenswrapper[4980]: I1206 03:35:32.834806 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:32 crc kubenswrapper[4980]: I1206 03:35:32.834838 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:32 crc kubenswrapper[4980]: I1206 03:35:32.834866 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:32Z","lastTransitionTime":"2025-12-06T03:35:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:32 crc kubenswrapper[4980]: I1206 03:35:32.938184 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:32 crc kubenswrapper[4980]: I1206 03:35:32.938238 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:32 crc kubenswrapper[4980]: I1206 03:35:32.938251 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:32 crc kubenswrapper[4980]: I1206 03:35:32.938269 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:32 crc kubenswrapper[4980]: I1206 03:35:32.938282 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:32Z","lastTransitionTime":"2025-12-06T03:35:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:33 crc kubenswrapper[4980]: I1206 03:35:33.041491 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:33 crc kubenswrapper[4980]: I1206 03:35:33.041612 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:33 crc kubenswrapper[4980]: I1206 03:35:33.041638 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:33 crc kubenswrapper[4980]: I1206 03:35:33.041665 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:33 crc kubenswrapper[4980]: I1206 03:35:33.041682 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:33Z","lastTransitionTime":"2025-12-06T03:35:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:33 crc kubenswrapper[4980]: I1206 03:35:33.145337 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:33 crc kubenswrapper[4980]: I1206 03:35:33.145412 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:33 crc kubenswrapper[4980]: I1206 03:35:33.145436 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:33 crc kubenswrapper[4980]: I1206 03:35:33.145461 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:33 crc kubenswrapper[4980]: I1206 03:35:33.145478 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:33Z","lastTransitionTime":"2025-12-06T03:35:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:33 crc kubenswrapper[4980]: I1206 03:35:33.184099 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 03:35:33 crc kubenswrapper[4980]: E1206 03:35:33.184288 4980 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 03:35:33 crc kubenswrapper[4980]: I1206 03:35:33.248397 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:33 crc kubenswrapper[4980]: I1206 03:35:33.248467 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:33 crc kubenswrapper[4980]: I1206 03:35:33.248484 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:33 crc kubenswrapper[4980]: I1206 03:35:33.248509 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:33 crc kubenswrapper[4980]: I1206 03:35:33.248559 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:33Z","lastTransitionTime":"2025-12-06T03:35:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:33 crc kubenswrapper[4980]: I1206 03:35:33.351479 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:33 crc kubenswrapper[4980]: I1206 03:35:33.351623 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:33 crc kubenswrapper[4980]: I1206 03:35:33.351653 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:33 crc kubenswrapper[4980]: I1206 03:35:33.351688 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:33 crc kubenswrapper[4980]: I1206 03:35:33.351712 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:33Z","lastTransitionTime":"2025-12-06T03:35:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:33 crc kubenswrapper[4980]: I1206 03:35:33.454972 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:33 crc kubenswrapper[4980]: I1206 03:35:33.455047 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:33 crc kubenswrapper[4980]: I1206 03:35:33.455070 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:33 crc kubenswrapper[4980]: I1206 03:35:33.455100 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:33 crc kubenswrapper[4980]: I1206 03:35:33.455121 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:33Z","lastTransitionTime":"2025-12-06T03:35:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:33 crc kubenswrapper[4980]: I1206 03:35:33.558462 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:33 crc kubenswrapper[4980]: I1206 03:35:33.558558 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:33 crc kubenswrapper[4980]: I1206 03:35:33.558588 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:33 crc kubenswrapper[4980]: I1206 03:35:33.558618 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:33 crc kubenswrapper[4980]: I1206 03:35:33.558639 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:33Z","lastTransitionTime":"2025-12-06T03:35:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:33 crc kubenswrapper[4980]: I1206 03:35:33.661621 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:33 crc kubenswrapper[4980]: I1206 03:35:33.661747 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:33 crc kubenswrapper[4980]: I1206 03:35:33.661771 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:33 crc kubenswrapper[4980]: I1206 03:35:33.661800 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:33 crc kubenswrapper[4980]: I1206 03:35:33.661823 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:33Z","lastTransitionTime":"2025-12-06T03:35:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:33 crc kubenswrapper[4980]: I1206 03:35:33.764712 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:33 crc kubenswrapper[4980]: I1206 03:35:33.764744 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:33 crc kubenswrapper[4980]: I1206 03:35:33.764752 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:33 crc kubenswrapper[4980]: I1206 03:35:33.764766 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:33 crc kubenswrapper[4980]: I1206 03:35:33.764774 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:33Z","lastTransitionTime":"2025-12-06T03:35:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:33 crc kubenswrapper[4980]: I1206 03:35:33.868102 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:33 crc kubenswrapper[4980]: I1206 03:35:33.868140 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:33 crc kubenswrapper[4980]: I1206 03:35:33.868156 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:33 crc kubenswrapper[4980]: I1206 03:35:33.868175 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:33 crc kubenswrapper[4980]: I1206 03:35:33.868188 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:33Z","lastTransitionTime":"2025-12-06T03:35:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:33 crc kubenswrapper[4980]: I1206 03:35:33.971058 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:33 crc kubenswrapper[4980]: I1206 03:35:33.971093 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:33 crc kubenswrapper[4980]: I1206 03:35:33.971105 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:33 crc kubenswrapper[4980]: I1206 03:35:33.971122 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:33 crc kubenswrapper[4980]: I1206 03:35:33.971130 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:33Z","lastTransitionTime":"2025-12-06T03:35:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:34 crc kubenswrapper[4980]: I1206 03:35:34.074138 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:34 crc kubenswrapper[4980]: I1206 03:35:34.074207 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:34 crc kubenswrapper[4980]: I1206 03:35:34.074235 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:34 crc kubenswrapper[4980]: I1206 03:35:34.074263 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:34 crc kubenswrapper[4980]: I1206 03:35:34.074279 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:34Z","lastTransitionTime":"2025-12-06T03:35:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:34 crc kubenswrapper[4980]: I1206 03:35:34.177280 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:34 crc kubenswrapper[4980]: I1206 03:35:34.177337 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:34 crc kubenswrapper[4980]: I1206 03:35:34.177355 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:34 crc kubenswrapper[4980]: I1206 03:35:34.177379 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:34 crc kubenswrapper[4980]: I1206 03:35:34.177397 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:34Z","lastTransitionTime":"2025-12-06T03:35:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:34 crc kubenswrapper[4980]: I1206 03:35:34.184016 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 03:35:34 crc kubenswrapper[4980]: I1206 03:35:34.184064 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 03:35:34 crc kubenswrapper[4980]: E1206 03:35:34.184180 4980 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 03:35:34 crc kubenswrapper[4980]: I1206 03:35:34.184245 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-mcprh" Dec 06 03:35:34 crc kubenswrapper[4980]: E1206 03:35:34.184736 4980 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-mcprh" podUID="e6e88c5e-a255-449a-ae38-ac4d73a8e984" Dec 06 03:35:34 crc kubenswrapper[4980]: E1206 03:35:34.184908 4980 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 03:35:34 crc kubenswrapper[4980]: I1206 03:35:34.185233 4980 scope.go:117] "RemoveContainer" containerID="33fa9bd73b9af5df6de4e7ed019a512334f465b830a3f4396cde14989658bf55" Dec 06 03:35:34 crc kubenswrapper[4980]: E1206 03:35:34.185479 4980 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-5w4fr_openshift-ovn-kubernetes(5af0e768-3c9c-4401-ab56-57516cd3170e)\"" pod="openshift-ovn-kubernetes/ovnkube-node-5w4fr" podUID="5af0e768-3c9c-4401-ab56-57516cd3170e" Dec 06 03:35:34 crc kubenswrapper[4980]: I1206 03:35:34.279614 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:34 crc kubenswrapper[4980]: I1206 03:35:34.279692 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:34 crc kubenswrapper[4980]: I1206 03:35:34.279718 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:34 crc kubenswrapper[4980]: I1206 03:35:34.279741 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:34 crc kubenswrapper[4980]: I1206 03:35:34.279761 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:34Z","lastTransitionTime":"2025-12-06T03:35:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:34 crc kubenswrapper[4980]: I1206 03:35:34.383198 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:34 crc kubenswrapper[4980]: I1206 03:35:34.383276 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:34 crc kubenswrapper[4980]: I1206 03:35:34.383299 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:34 crc kubenswrapper[4980]: I1206 03:35:34.383332 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:34 crc kubenswrapper[4980]: I1206 03:35:34.383356 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:34Z","lastTransitionTime":"2025-12-06T03:35:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:34 crc kubenswrapper[4980]: I1206 03:35:34.486080 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:34 crc kubenswrapper[4980]: I1206 03:35:34.486184 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:34 crc kubenswrapper[4980]: I1206 03:35:34.486202 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:34 crc kubenswrapper[4980]: I1206 03:35:34.486229 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:34 crc kubenswrapper[4980]: I1206 03:35:34.486254 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:34Z","lastTransitionTime":"2025-12-06T03:35:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:34 crc kubenswrapper[4980]: I1206 03:35:34.590086 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:34 crc kubenswrapper[4980]: I1206 03:35:34.590164 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:34 crc kubenswrapper[4980]: I1206 03:35:34.590186 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:34 crc kubenswrapper[4980]: I1206 03:35:34.590212 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:34 crc kubenswrapper[4980]: I1206 03:35:34.590230 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:34Z","lastTransitionTime":"2025-12-06T03:35:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:34 crc kubenswrapper[4980]: I1206 03:35:34.693259 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:34 crc kubenswrapper[4980]: I1206 03:35:34.693315 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:34 crc kubenswrapper[4980]: I1206 03:35:34.693326 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:34 crc kubenswrapper[4980]: I1206 03:35:34.693347 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:34 crc kubenswrapper[4980]: I1206 03:35:34.693367 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:34Z","lastTransitionTime":"2025-12-06T03:35:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:34 crc kubenswrapper[4980]: I1206 03:35:34.796050 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:34 crc kubenswrapper[4980]: I1206 03:35:34.796088 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:34 crc kubenswrapper[4980]: I1206 03:35:34.796096 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:34 crc kubenswrapper[4980]: I1206 03:35:34.796111 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:34 crc kubenswrapper[4980]: I1206 03:35:34.796122 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:34Z","lastTransitionTime":"2025-12-06T03:35:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:34 crc kubenswrapper[4980]: I1206 03:35:34.901780 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:34 crc kubenswrapper[4980]: I1206 03:35:34.901912 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:34 crc kubenswrapper[4980]: I1206 03:35:34.901927 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:34 crc kubenswrapper[4980]: I1206 03:35:34.901947 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:34 crc kubenswrapper[4980]: I1206 03:35:34.901967 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:34Z","lastTransitionTime":"2025-12-06T03:35:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:35 crc kubenswrapper[4980]: I1206 03:35:35.007191 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:35 crc kubenswrapper[4980]: I1206 03:35:35.007256 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:35 crc kubenswrapper[4980]: I1206 03:35:35.007272 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:35 crc kubenswrapper[4980]: I1206 03:35:35.007296 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:35 crc kubenswrapper[4980]: I1206 03:35:35.007312 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:35Z","lastTransitionTime":"2025-12-06T03:35:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:35 crc kubenswrapper[4980]: I1206 03:35:35.110374 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:35 crc kubenswrapper[4980]: I1206 03:35:35.110430 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:35 crc kubenswrapper[4980]: I1206 03:35:35.110444 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:35 crc kubenswrapper[4980]: I1206 03:35:35.110462 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:35 crc kubenswrapper[4980]: I1206 03:35:35.110474 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:35Z","lastTransitionTime":"2025-12-06T03:35:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:35 crc kubenswrapper[4980]: I1206 03:35:35.184020 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 03:35:35 crc kubenswrapper[4980]: E1206 03:35:35.184235 4980 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 03:35:35 crc kubenswrapper[4980]: I1206 03:35:35.212929 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:35 crc kubenswrapper[4980]: I1206 03:35:35.212986 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:35 crc kubenswrapper[4980]: I1206 03:35:35.212998 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:35 crc kubenswrapper[4980]: I1206 03:35:35.213016 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:35 crc kubenswrapper[4980]: I1206 03:35:35.213031 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:35Z","lastTransitionTime":"2025-12-06T03:35:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:35 crc kubenswrapper[4980]: I1206 03:35:35.315991 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:35 crc kubenswrapper[4980]: I1206 03:35:35.316048 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:35 crc kubenswrapper[4980]: I1206 03:35:35.316060 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:35 crc kubenswrapper[4980]: I1206 03:35:35.316078 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:35 crc kubenswrapper[4980]: I1206 03:35:35.316094 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:35Z","lastTransitionTime":"2025-12-06T03:35:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:35 crc kubenswrapper[4980]: I1206 03:35:35.419334 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:35 crc kubenswrapper[4980]: I1206 03:35:35.419395 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:35 crc kubenswrapper[4980]: I1206 03:35:35.419422 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:35 crc kubenswrapper[4980]: I1206 03:35:35.419455 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:35 crc kubenswrapper[4980]: I1206 03:35:35.419479 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:35Z","lastTransitionTime":"2025-12-06T03:35:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:35 crc kubenswrapper[4980]: I1206 03:35:35.522690 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:35 crc kubenswrapper[4980]: I1206 03:35:35.522791 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:35 crc kubenswrapper[4980]: I1206 03:35:35.522818 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:35 crc kubenswrapper[4980]: I1206 03:35:35.522848 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:35 crc kubenswrapper[4980]: I1206 03:35:35.522871 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:35Z","lastTransitionTime":"2025-12-06T03:35:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:35 crc kubenswrapper[4980]: I1206 03:35:35.625848 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:35 crc kubenswrapper[4980]: I1206 03:35:35.625917 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:35 crc kubenswrapper[4980]: I1206 03:35:35.625939 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:35 crc kubenswrapper[4980]: I1206 03:35:35.625967 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:35 crc kubenswrapper[4980]: I1206 03:35:35.625984 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:35Z","lastTransitionTime":"2025-12-06T03:35:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:35 crc kubenswrapper[4980]: I1206 03:35:35.728868 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:35 crc kubenswrapper[4980]: I1206 03:35:35.728999 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:35 crc kubenswrapper[4980]: I1206 03:35:35.729019 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:35 crc kubenswrapper[4980]: I1206 03:35:35.729044 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:35 crc kubenswrapper[4980]: I1206 03:35:35.729091 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:35Z","lastTransitionTime":"2025-12-06T03:35:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:35 crc kubenswrapper[4980]: I1206 03:35:35.832925 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:35 crc kubenswrapper[4980]: I1206 03:35:35.832977 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:35 crc kubenswrapper[4980]: I1206 03:35:35.832989 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:35 crc kubenswrapper[4980]: I1206 03:35:35.833027 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:35 crc kubenswrapper[4980]: I1206 03:35:35.833056 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:35Z","lastTransitionTime":"2025-12-06T03:35:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:35 crc kubenswrapper[4980]: I1206 03:35:35.936825 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:35 crc kubenswrapper[4980]: I1206 03:35:35.936867 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:35 crc kubenswrapper[4980]: I1206 03:35:35.936878 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:35 crc kubenswrapper[4980]: I1206 03:35:35.936897 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:35 crc kubenswrapper[4980]: I1206 03:35:35.936908 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:35Z","lastTransitionTime":"2025-12-06T03:35:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:36 crc kubenswrapper[4980]: I1206 03:35:36.040805 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:36 crc kubenswrapper[4980]: I1206 03:35:36.040846 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:36 crc kubenswrapper[4980]: I1206 03:35:36.040859 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:36 crc kubenswrapper[4980]: I1206 03:35:36.040878 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:36 crc kubenswrapper[4980]: I1206 03:35:36.040891 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:36Z","lastTransitionTime":"2025-12-06T03:35:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:36 crc kubenswrapper[4980]: I1206 03:35:36.143304 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:36 crc kubenswrapper[4980]: I1206 03:35:36.143350 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:36 crc kubenswrapper[4980]: I1206 03:35:36.143362 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:36 crc kubenswrapper[4980]: I1206 03:35:36.143380 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:36 crc kubenswrapper[4980]: I1206 03:35:36.143393 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:36Z","lastTransitionTime":"2025-12-06T03:35:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:36 crc kubenswrapper[4980]: I1206 03:35:36.184136 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 03:35:36 crc kubenswrapper[4980]: I1206 03:35:36.184225 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 03:35:36 crc kubenswrapper[4980]: I1206 03:35:36.184235 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-mcprh" Dec 06 03:35:36 crc kubenswrapper[4980]: E1206 03:35:36.184672 4980 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 03:35:36 crc kubenswrapper[4980]: E1206 03:35:36.184840 4980 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 03:35:36 crc kubenswrapper[4980]: E1206 03:35:36.184934 4980 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-mcprh" podUID="e6e88c5e-a255-449a-ae38-ac4d73a8e984" Dec 06 03:35:36 crc kubenswrapper[4980]: I1206 03:35:36.246273 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:36 crc kubenswrapper[4980]: I1206 03:35:36.246317 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:36 crc kubenswrapper[4980]: I1206 03:35:36.246326 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:36 crc kubenswrapper[4980]: I1206 03:35:36.246341 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:36 crc kubenswrapper[4980]: I1206 03:35:36.246352 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:36Z","lastTransitionTime":"2025-12-06T03:35:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:36 crc kubenswrapper[4980]: I1206 03:35:36.349563 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:36 crc kubenswrapper[4980]: I1206 03:35:36.349647 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:36 crc kubenswrapper[4980]: I1206 03:35:36.349669 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:36 crc kubenswrapper[4980]: I1206 03:35:36.349702 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:36 crc kubenswrapper[4980]: I1206 03:35:36.349727 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:36Z","lastTransitionTime":"2025-12-06T03:35:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:36 crc kubenswrapper[4980]: I1206 03:35:36.452637 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:36 crc kubenswrapper[4980]: I1206 03:35:36.452686 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:36 crc kubenswrapper[4980]: I1206 03:35:36.452697 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:36 crc kubenswrapper[4980]: I1206 03:35:36.452714 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:36 crc kubenswrapper[4980]: I1206 03:35:36.452726 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:36Z","lastTransitionTime":"2025-12-06T03:35:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:36 crc kubenswrapper[4980]: I1206 03:35:36.555969 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:36 crc kubenswrapper[4980]: I1206 03:35:36.556067 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:36 crc kubenswrapper[4980]: I1206 03:35:36.556087 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:36 crc kubenswrapper[4980]: I1206 03:35:36.556113 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:36 crc kubenswrapper[4980]: I1206 03:35:36.556130 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:36Z","lastTransitionTime":"2025-12-06T03:35:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:36 crc kubenswrapper[4980]: I1206 03:35:36.659820 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:36 crc kubenswrapper[4980]: I1206 03:35:36.659929 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:36 crc kubenswrapper[4980]: I1206 03:35:36.659948 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:36 crc kubenswrapper[4980]: I1206 03:35:36.659973 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:36 crc kubenswrapper[4980]: I1206 03:35:36.659989 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:36Z","lastTransitionTime":"2025-12-06T03:35:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:36 crc kubenswrapper[4980]: I1206 03:35:36.762662 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:36 crc kubenswrapper[4980]: I1206 03:35:36.762732 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:36 crc kubenswrapper[4980]: I1206 03:35:36.762757 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:36 crc kubenswrapper[4980]: I1206 03:35:36.762788 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:36 crc kubenswrapper[4980]: I1206 03:35:36.762811 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:36Z","lastTransitionTime":"2025-12-06T03:35:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:36 crc kubenswrapper[4980]: I1206 03:35:36.867789 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:36 crc kubenswrapper[4980]: I1206 03:35:36.867873 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:36 crc kubenswrapper[4980]: I1206 03:35:36.867935 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:36 crc kubenswrapper[4980]: I1206 03:35:36.867967 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:36 crc kubenswrapper[4980]: I1206 03:35:36.867988 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:36Z","lastTransitionTime":"2025-12-06T03:35:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:36 crc kubenswrapper[4980]: I1206 03:35:36.971402 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:36 crc kubenswrapper[4980]: I1206 03:35:36.971470 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:36 crc kubenswrapper[4980]: I1206 03:35:36.971488 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:36 crc kubenswrapper[4980]: I1206 03:35:36.971544 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:36 crc kubenswrapper[4980]: I1206 03:35:36.971565 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:36Z","lastTransitionTime":"2025-12-06T03:35:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:37 crc kubenswrapper[4980]: I1206 03:35:37.074453 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:37 crc kubenswrapper[4980]: I1206 03:35:37.074548 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:37 crc kubenswrapper[4980]: I1206 03:35:37.074563 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:37 crc kubenswrapper[4980]: I1206 03:35:37.074580 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:37 crc kubenswrapper[4980]: I1206 03:35:37.074593 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:37Z","lastTransitionTime":"2025-12-06T03:35:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:37 crc kubenswrapper[4980]: I1206 03:35:37.177871 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:37 crc kubenswrapper[4980]: I1206 03:35:37.177951 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:37 crc kubenswrapper[4980]: I1206 03:35:37.177971 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:37 crc kubenswrapper[4980]: I1206 03:35:37.177997 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:37 crc kubenswrapper[4980]: I1206 03:35:37.178014 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:37Z","lastTransitionTime":"2025-12-06T03:35:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:37 crc kubenswrapper[4980]: I1206 03:35:37.184287 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 03:35:37 crc kubenswrapper[4980]: E1206 03:35:37.184435 4980 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 03:35:37 crc kubenswrapper[4980]: I1206 03:35:37.280060 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:37 crc kubenswrapper[4980]: I1206 03:35:37.280093 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:37 crc kubenswrapper[4980]: I1206 03:35:37.280100 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:37 crc kubenswrapper[4980]: I1206 03:35:37.280114 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:37 crc kubenswrapper[4980]: I1206 03:35:37.280123 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:37Z","lastTransitionTime":"2025-12-06T03:35:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:37 crc kubenswrapper[4980]: I1206 03:35:37.382212 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:37 crc kubenswrapper[4980]: I1206 03:35:37.382280 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:37 crc kubenswrapper[4980]: I1206 03:35:37.382291 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:37 crc kubenswrapper[4980]: I1206 03:35:37.382308 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:37 crc kubenswrapper[4980]: I1206 03:35:37.382320 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:37Z","lastTransitionTime":"2025-12-06T03:35:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:37 crc kubenswrapper[4980]: I1206 03:35:37.485559 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:37 crc kubenswrapper[4980]: I1206 03:35:37.485637 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:37 crc kubenswrapper[4980]: I1206 03:35:37.485656 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:37 crc kubenswrapper[4980]: I1206 03:35:37.485679 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:37 crc kubenswrapper[4980]: I1206 03:35:37.485697 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:37Z","lastTransitionTime":"2025-12-06T03:35:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:37 crc kubenswrapper[4980]: I1206 03:35:37.589549 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:37 crc kubenswrapper[4980]: I1206 03:35:37.589611 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:37 crc kubenswrapper[4980]: I1206 03:35:37.589633 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:37 crc kubenswrapper[4980]: I1206 03:35:37.589662 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:37 crc kubenswrapper[4980]: I1206 03:35:37.589682 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:37Z","lastTransitionTime":"2025-12-06T03:35:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:37 crc kubenswrapper[4980]: I1206 03:35:37.692898 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:37 crc kubenswrapper[4980]: I1206 03:35:37.692948 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:37 crc kubenswrapper[4980]: I1206 03:35:37.692964 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:37 crc kubenswrapper[4980]: I1206 03:35:37.692988 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:37 crc kubenswrapper[4980]: I1206 03:35:37.693006 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:37Z","lastTransitionTime":"2025-12-06T03:35:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:37 crc kubenswrapper[4980]: I1206 03:35:37.796206 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:37 crc kubenswrapper[4980]: I1206 03:35:37.796284 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:37 crc kubenswrapper[4980]: I1206 03:35:37.796311 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:37 crc kubenswrapper[4980]: I1206 03:35:37.796341 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:37 crc kubenswrapper[4980]: I1206 03:35:37.796363 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:37Z","lastTransitionTime":"2025-12-06T03:35:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:37 crc kubenswrapper[4980]: I1206 03:35:37.900666 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:37 crc kubenswrapper[4980]: I1206 03:35:37.900730 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:37 crc kubenswrapper[4980]: I1206 03:35:37.900748 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:37 crc kubenswrapper[4980]: I1206 03:35:37.900774 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:37 crc kubenswrapper[4980]: I1206 03:35:37.900790 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:37Z","lastTransitionTime":"2025-12-06T03:35:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:38 crc kubenswrapper[4980]: I1206 03:35:38.007097 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:38 crc kubenswrapper[4980]: I1206 03:35:38.007214 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:38 crc kubenswrapper[4980]: I1206 03:35:38.007235 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:38 crc kubenswrapper[4980]: I1206 03:35:38.007302 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:38 crc kubenswrapper[4980]: I1206 03:35:38.007324 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:38Z","lastTransitionTime":"2025-12-06T03:35:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:38 crc kubenswrapper[4980]: I1206 03:35:38.110852 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:38 crc kubenswrapper[4980]: I1206 03:35:38.110942 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:38 crc kubenswrapper[4980]: I1206 03:35:38.110966 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:38 crc kubenswrapper[4980]: I1206 03:35:38.110996 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:38 crc kubenswrapper[4980]: I1206 03:35:38.111016 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:38Z","lastTransitionTime":"2025-12-06T03:35:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:38 crc kubenswrapper[4980]: I1206 03:35:38.183712 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 03:35:38 crc kubenswrapper[4980]: I1206 03:35:38.183750 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 03:35:38 crc kubenswrapper[4980]: I1206 03:35:38.183767 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-mcprh" Dec 06 03:35:38 crc kubenswrapper[4980]: E1206 03:35:38.184011 4980 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 03:35:38 crc kubenswrapper[4980]: E1206 03:35:38.184110 4980 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 03:35:38 crc kubenswrapper[4980]: E1206 03:35:38.184284 4980 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-mcprh" podUID="e6e88c5e-a255-449a-ae38-ac4d73a8e984" Dec 06 03:35:38 crc kubenswrapper[4980]: I1206 03:35:38.217603 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:38 crc kubenswrapper[4980]: I1206 03:35:38.217682 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:38 crc kubenswrapper[4980]: I1206 03:35:38.217710 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:38 crc kubenswrapper[4980]: I1206 03:35:38.217744 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:38 crc kubenswrapper[4980]: I1206 03:35:38.217767 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:38Z","lastTransitionTime":"2025-12-06T03:35:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:38 crc kubenswrapper[4980]: I1206 03:35:38.321382 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:38 crc kubenswrapper[4980]: I1206 03:35:38.321433 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:38 crc kubenswrapper[4980]: I1206 03:35:38.321449 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:38 crc kubenswrapper[4980]: I1206 03:35:38.321473 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:38 crc kubenswrapper[4980]: I1206 03:35:38.321553 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:38Z","lastTransitionTime":"2025-12-06T03:35:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:38 crc kubenswrapper[4980]: I1206 03:35:38.424188 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:38 crc kubenswrapper[4980]: I1206 03:35:38.424224 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:38 crc kubenswrapper[4980]: I1206 03:35:38.424236 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:38 crc kubenswrapper[4980]: I1206 03:35:38.424251 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:38 crc kubenswrapper[4980]: I1206 03:35:38.424261 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:38Z","lastTransitionTime":"2025-12-06T03:35:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:38 crc kubenswrapper[4980]: I1206 03:35:38.526968 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:38 crc kubenswrapper[4980]: I1206 03:35:38.527025 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:38 crc kubenswrapper[4980]: I1206 03:35:38.527045 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:38 crc kubenswrapper[4980]: I1206 03:35:38.527075 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:38 crc kubenswrapper[4980]: I1206 03:35:38.527095 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:38Z","lastTransitionTime":"2025-12-06T03:35:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:38 crc kubenswrapper[4980]: I1206 03:35:38.630818 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:38 crc kubenswrapper[4980]: I1206 03:35:38.630898 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:38 crc kubenswrapper[4980]: I1206 03:35:38.630923 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:38 crc kubenswrapper[4980]: I1206 03:35:38.630955 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:38 crc kubenswrapper[4980]: I1206 03:35:38.630977 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:38Z","lastTransitionTime":"2025-12-06T03:35:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:38 crc kubenswrapper[4980]: I1206 03:35:38.734547 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:38 crc kubenswrapper[4980]: I1206 03:35:38.734592 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:38 crc kubenswrapper[4980]: I1206 03:35:38.734602 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:38 crc kubenswrapper[4980]: I1206 03:35:38.734617 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:38 crc kubenswrapper[4980]: I1206 03:35:38.734629 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:38Z","lastTransitionTime":"2025-12-06T03:35:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:38 crc kubenswrapper[4980]: I1206 03:35:38.837707 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:38 crc kubenswrapper[4980]: I1206 03:35:38.837761 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:38 crc kubenswrapper[4980]: I1206 03:35:38.837774 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:38 crc kubenswrapper[4980]: I1206 03:35:38.837791 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:38 crc kubenswrapper[4980]: I1206 03:35:38.837805 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:38Z","lastTransitionTime":"2025-12-06T03:35:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:38 crc kubenswrapper[4980]: I1206 03:35:38.941204 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:38 crc kubenswrapper[4980]: I1206 03:35:38.941274 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:38 crc kubenswrapper[4980]: I1206 03:35:38.941287 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:38 crc kubenswrapper[4980]: I1206 03:35:38.941307 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:38 crc kubenswrapper[4980]: I1206 03:35:38.941320 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:38Z","lastTransitionTime":"2025-12-06T03:35:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:39 crc kubenswrapper[4980]: I1206 03:35:39.044742 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:39 crc kubenswrapper[4980]: I1206 03:35:39.044796 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:39 crc kubenswrapper[4980]: I1206 03:35:39.044811 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:39 crc kubenswrapper[4980]: I1206 03:35:39.044829 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:39 crc kubenswrapper[4980]: I1206 03:35:39.044857 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:39Z","lastTransitionTime":"2025-12-06T03:35:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:39 crc kubenswrapper[4980]: I1206 03:35:39.147090 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:39 crc kubenswrapper[4980]: I1206 03:35:39.147137 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:39 crc kubenswrapper[4980]: I1206 03:35:39.147146 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:39 crc kubenswrapper[4980]: I1206 03:35:39.147161 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:39 crc kubenswrapper[4980]: I1206 03:35:39.147171 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:39Z","lastTransitionTime":"2025-12-06T03:35:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:39 crc kubenswrapper[4980]: I1206 03:35:39.183735 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 03:35:39 crc kubenswrapper[4980]: E1206 03:35:39.183987 4980 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 03:35:39 crc kubenswrapper[4980]: I1206 03:35:39.249475 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:39 crc kubenswrapper[4980]: I1206 03:35:39.249542 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:39 crc kubenswrapper[4980]: I1206 03:35:39.249559 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:39 crc kubenswrapper[4980]: I1206 03:35:39.249582 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:39 crc kubenswrapper[4980]: I1206 03:35:39.249597 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:39Z","lastTransitionTime":"2025-12-06T03:35:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:39 crc kubenswrapper[4980]: I1206 03:35:39.353501 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:39 crc kubenswrapper[4980]: I1206 03:35:39.353573 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:39 crc kubenswrapper[4980]: I1206 03:35:39.353582 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:39 crc kubenswrapper[4980]: I1206 03:35:39.353600 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:39 crc kubenswrapper[4980]: I1206 03:35:39.353630 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:39Z","lastTransitionTime":"2025-12-06T03:35:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:39 crc kubenswrapper[4980]: I1206 03:35:39.457161 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:39 crc kubenswrapper[4980]: I1206 03:35:39.457204 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:39 crc kubenswrapper[4980]: I1206 03:35:39.457214 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:39 crc kubenswrapper[4980]: I1206 03:35:39.457230 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:39 crc kubenswrapper[4980]: I1206 03:35:39.457243 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:39Z","lastTransitionTime":"2025-12-06T03:35:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:39 crc kubenswrapper[4980]: I1206 03:35:39.560827 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:39 crc kubenswrapper[4980]: I1206 03:35:39.560873 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:39 crc kubenswrapper[4980]: I1206 03:35:39.560882 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:39 crc kubenswrapper[4980]: I1206 03:35:39.560909 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:39 crc kubenswrapper[4980]: I1206 03:35:39.560919 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:39Z","lastTransitionTime":"2025-12-06T03:35:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:39 crc kubenswrapper[4980]: I1206 03:35:39.663343 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:39 crc kubenswrapper[4980]: I1206 03:35:39.663386 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:39 crc kubenswrapper[4980]: I1206 03:35:39.663402 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:39 crc kubenswrapper[4980]: I1206 03:35:39.663421 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:39 crc kubenswrapper[4980]: I1206 03:35:39.663434 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:39Z","lastTransitionTime":"2025-12-06T03:35:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:39 crc kubenswrapper[4980]: I1206 03:35:39.766467 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:39 crc kubenswrapper[4980]: I1206 03:35:39.766570 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:39 crc kubenswrapper[4980]: I1206 03:35:39.766594 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:39 crc kubenswrapper[4980]: I1206 03:35:39.766628 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:39 crc kubenswrapper[4980]: I1206 03:35:39.766650 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:39Z","lastTransitionTime":"2025-12-06T03:35:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:39 crc kubenswrapper[4980]: I1206 03:35:39.870230 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:39 crc kubenswrapper[4980]: I1206 03:35:39.870288 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:39 crc kubenswrapper[4980]: I1206 03:35:39.870305 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:39 crc kubenswrapper[4980]: I1206 03:35:39.870328 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:39 crc kubenswrapper[4980]: I1206 03:35:39.870345 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:39Z","lastTransitionTime":"2025-12-06T03:35:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:39 crc kubenswrapper[4980]: I1206 03:35:39.973334 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:39 crc kubenswrapper[4980]: I1206 03:35:39.973433 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:39 crc kubenswrapper[4980]: I1206 03:35:39.973470 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:39 crc kubenswrapper[4980]: I1206 03:35:39.973499 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:39 crc kubenswrapper[4980]: I1206 03:35:39.973547 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:39Z","lastTransitionTime":"2025-12-06T03:35:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:40 crc kubenswrapper[4980]: I1206 03:35:40.077733 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:40 crc kubenswrapper[4980]: I1206 03:35:40.077777 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:40 crc kubenswrapper[4980]: I1206 03:35:40.077798 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:40 crc kubenswrapper[4980]: I1206 03:35:40.077817 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:40 crc kubenswrapper[4980]: I1206 03:35:40.077830 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:40Z","lastTransitionTime":"2025-12-06T03:35:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:40 crc kubenswrapper[4980]: I1206 03:35:40.180705 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:40 crc kubenswrapper[4980]: I1206 03:35:40.180817 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:40 crc kubenswrapper[4980]: I1206 03:35:40.180835 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:40 crc kubenswrapper[4980]: I1206 03:35:40.180865 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:40 crc kubenswrapper[4980]: I1206 03:35:40.180883 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:40Z","lastTransitionTime":"2025-12-06T03:35:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:40 crc kubenswrapper[4980]: I1206 03:35:40.184084 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 03:35:40 crc kubenswrapper[4980]: E1206 03:35:40.184272 4980 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 03:35:40 crc kubenswrapper[4980]: I1206 03:35:40.184395 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 03:35:40 crc kubenswrapper[4980]: I1206 03:35:40.184473 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-mcprh" Dec 06 03:35:40 crc kubenswrapper[4980]: E1206 03:35:40.184651 4980 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-mcprh" podUID="e6e88c5e-a255-449a-ae38-ac4d73a8e984" Dec 06 03:35:40 crc kubenswrapper[4980]: E1206 03:35:40.184819 4980 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 03:35:40 crc kubenswrapper[4980]: I1206 03:35:40.284612 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:40 crc kubenswrapper[4980]: I1206 03:35:40.284675 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:40 crc kubenswrapper[4980]: I1206 03:35:40.284693 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:40 crc kubenswrapper[4980]: I1206 03:35:40.284718 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:40 crc kubenswrapper[4980]: I1206 03:35:40.284736 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:40Z","lastTransitionTime":"2025-12-06T03:35:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:40 crc kubenswrapper[4980]: I1206 03:35:40.391610 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:40 crc kubenswrapper[4980]: I1206 03:35:40.391688 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:40 crc kubenswrapper[4980]: I1206 03:35:40.391701 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:40 crc kubenswrapper[4980]: I1206 03:35:40.392044 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:40 crc kubenswrapper[4980]: I1206 03:35:40.392310 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:40Z","lastTransitionTime":"2025-12-06T03:35:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:40 crc kubenswrapper[4980]: I1206 03:35:40.427085 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 03:35:40 crc kubenswrapper[4980]: I1206 03:35:40.427343 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 03:35:40 crc kubenswrapper[4980]: I1206 03:35:40.427408 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 03:35:40 crc kubenswrapper[4980]: I1206 03:35:40.427627 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 03:35:40 crc kubenswrapper[4980]: I1206 03:35:40.427697 4980 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T03:35:40Z","lastTransitionTime":"2025-12-06T03:35:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 03:35:40 crc kubenswrapper[4980]: I1206 03:35:40.501092 4980 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-version/cluster-version-operator-5c965bbfc6-prc2s"] Dec 06 03:35:40 crc kubenswrapper[4980]: I1206 03:35:40.502343 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-prc2s" Dec 06 03:35:40 crc kubenswrapper[4980]: I1206 03:35:40.506079 4980 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Dec 06 03:35:40 crc kubenswrapper[4980]: I1206 03:35:40.507265 4980 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Dec 06 03:35:40 crc kubenswrapper[4980]: I1206 03:35:40.507265 4980 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Dec 06 03:35:40 crc kubenswrapper[4980]: I1206 03:35:40.507310 4980 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Dec 06 03:35:40 crc kubenswrapper[4980]: I1206 03:35:40.543390 4980 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" podStartSLOduration=44.543340692 podStartE2EDuration="44.543340692s" podCreationTimestamp="2025-12-06 03:34:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 03:35:40.525987251 +0000 UTC m=+99.770738542" watchObservedRunningTime="2025-12-06 03:35:40.543340692 +0000 UTC m=+99.788091983" Dec 06 03:35:40 crc kubenswrapper[4980]: I1206 03:35:40.548197 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/82417d46-29da-4c49-8771-534a90f57c95-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-prc2s\" (UID: \"82417d46-29da-4c49-8771-534a90f57c95\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-prc2s" Dec 06 03:35:40 crc kubenswrapper[4980]: I1206 03:35:40.548247 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/82417d46-29da-4c49-8771-534a90f57c95-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-prc2s\" (UID: \"82417d46-29da-4c49-8771-534a90f57c95\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-prc2s" Dec 06 03:35:40 crc kubenswrapper[4980]: I1206 03:35:40.548282 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/82417d46-29da-4c49-8771-534a90f57c95-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-prc2s\" (UID: \"82417d46-29da-4c49-8771-534a90f57c95\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-prc2s" Dec 06 03:35:40 crc kubenswrapper[4980]: I1206 03:35:40.548314 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/82417d46-29da-4c49-8771-534a90f57c95-service-ca\") pod \"cluster-version-operator-5c965bbfc6-prc2s\" (UID: \"82417d46-29da-4c49-8771-534a90f57c95\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-prc2s" Dec 06 03:35:40 crc kubenswrapper[4980]: I1206 03:35:40.548338 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/82417d46-29da-4c49-8771-534a90f57c95-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-prc2s\" (UID: \"82417d46-29da-4c49-8771-534a90f57c95\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-prc2s" Dec 06 03:35:40 crc kubenswrapper[4980]: I1206 03:35:40.649802 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/82417d46-29da-4c49-8771-534a90f57c95-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-prc2s\" (UID: \"82417d46-29da-4c49-8771-534a90f57c95\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-prc2s" Dec 06 03:35:40 crc kubenswrapper[4980]: I1206 03:35:40.649865 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/82417d46-29da-4c49-8771-534a90f57c95-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-prc2s\" (UID: \"82417d46-29da-4c49-8771-534a90f57c95\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-prc2s" Dec 06 03:35:40 crc kubenswrapper[4980]: I1206 03:35:40.649915 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/82417d46-29da-4c49-8771-534a90f57c95-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-prc2s\" (UID: \"82417d46-29da-4c49-8771-534a90f57c95\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-prc2s" Dec 06 03:35:40 crc kubenswrapper[4980]: I1206 03:35:40.649936 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/82417d46-29da-4c49-8771-534a90f57c95-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-prc2s\" (UID: \"82417d46-29da-4c49-8771-534a90f57c95\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-prc2s" Dec 06 03:35:40 crc kubenswrapper[4980]: I1206 03:35:40.649960 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/82417d46-29da-4c49-8771-534a90f57c95-service-ca\") pod \"cluster-version-operator-5c965bbfc6-prc2s\" (UID: \"82417d46-29da-4c49-8771-534a90f57c95\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-prc2s" Dec 06 03:35:40 crc kubenswrapper[4980]: I1206 03:35:40.650014 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/82417d46-29da-4c49-8771-534a90f57c95-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-prc2s\" (UID: \"82417d46-29da-4c49-8771-534a90f57c95\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-prc2s" Dec 06 03:35:40 crc kubenswrapper[4980]: I1206 03:35:40.650077 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/82417d46-29da-4c49-8771-534a90f57c95-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-prc2s\" (UID: \"82417d46-29da-4c49-8771-534a90f57c95\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-prc2s" Dec 06 03:35:40 crc kubenswrapper[4980]: I1206 03:35:40.652193 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/82417d46-29da-4c49-8771-534a90f57c95-service-ca\") pod \"cluster-version-operator-5c965bbfc6-prc2s\" (UID: \"82417d46-29da-4c49-8771-534a90f57c95\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-prc2s" Dec 06 03:35:40 crc kubenswrapper[4980]: I1206 03:35:40.656694 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/82417d46-29da-4c49-8771-534a90f57c95-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-prc2s\" (UID: \"82417d46-29da-4c49-8771-534a90f57c95\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-prc2s" Dec 06 03:35:40 crc kubenswrapper[4980]: I1206 03:35:40.671007 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/82417d46-29da-4c49-8771-534a90f57c95-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-prc2s\" (UID: \"82417d46-29da-4c49-8771-534a90f57c95\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-prc2s" Dec 06 03:35:40 crc kubenswrapper[4980]: I1206 03:35:40.822450 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-prc2s" Dec 06 03:35:41 crc kubenswrapper[4980]: I1206 03:35:41.183899 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 03:35:41 crc kubenswrapper[4980]: E1206 03:35:41.186825 4980 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 03:35:41 crc kubenswrapper[4980]: I1206 03:35:41.238561 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-prc2s" event={"ID":"82417d46-29da-4c49-8771-534a90f57c95","Type":"ContainerStarted","Data":"9dfee423a76aa61efc7ee44b7199d25d992115effbf45910245db6ec61eb2779"} Dec 06 03:35:41 crc kubenswrapper[4980]: I1206 03:35:41.238631 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-prc2s" event={"ID":"82417d46-29da-4c49-8771-534a90f57c95","Type":"ContainerStarted","Data":"16236c3a04775a60f1c16445e3638d1f0f41fdb6ac43b7761569388651966949"} Dec 06 03:35:41 crc kubenswrapper[4980]: I1206 03:35:41.253010 4980 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-prc2s" podStartSLOduration=79.252984673 podStartE2EDuration="1m19.252984673s" podCreationTimestamp="2025-12-06 03:34:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 03:35:41.251989385 +0000 UTC m=+100.496740656" watchObservedRunningTime="2025-12-06 03:35:41.252984673 +0000 UTC m=+100.497735954" Dec 06 03:35:42 crc kubenswrapper[4980]: I1206 03:35:42.183774 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 03:35:42 crc kubenswrapper[4980]: I1206 03:35:42.183815 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 03:35:42 crc kubenswrapper[4980]: I1206 03:35:42.183934 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-mcprh" Dec 06 03:35:42 crc kubenswrapper[4980]: E1206 03:35:42.183958 4980 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 03:35:42 crc kubenswrapper[4980]: E1206 03:35:42.184045 4980 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 03:35:42 crc kubenswrapper[4980]: E1206 03:35:42.184239 4980 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-mcprh" podUID="e6e88c5e-a255-449a-ae38-ac4d73a8e984" Dec 06 03:35:42 crc kubenswrapper[4980]: I1206 03:35:42.569391 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/e6e88c5e-a255-449a-ae38-ac4d73a8e984-metrics-certs\") pod \"network-metrics-daemon-mcprh\" (UID: \"e6e88c5e-a255-449a-ae38-ac4d73a8e984\") " pod="openshift-multus/network-metrics-daemon-mcprh" Dec 06 03:35:42 crc kubenswrapper[4980]: E1206 03:35:42.569673 4980 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 06 03:35:42 crc kubenswrapper[4980]: E1206 03:35:42.569814 4980 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/e6e88c5e-a255-449a-ae38-ac4d73a8e984-metrics-certs podName:e6e88c5e-a255-449a-ae38-ac4d73a8e984 nodeName:}" failed. No retries permitted until 2025-12-06 03:36:46.569786706 +0000 UTC m=+165.814538017 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/e6e88c5e-a255-449a-ae38-ac4d73a8e984-metrics-certs") pod "network-metrics-daemon-mcprh" (UID: "e6e88c5e-a255-449a-ae38-ac4d73a8e984") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 06 03:35:43 crc kubenswrapper[4980]: I1206 03:35:43.183689 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 03:35:43 crc kubenswrapper[4980]: E1206 03:35:43.183975 4980 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 03:35:44 crc kubenswrapper[4980]: I1206 03:35:44.183807 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 03:35:44 crc kubenswrapper[4980]: I1206 03:35:44.183808 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-mcprh" Dec 06 03:35:44 crc kubenswrapper[4980]: I1206 03:35:44.183846 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 03:35:44 crc kubenswrapper[4980]: E1206 03:35:44.184001 4980 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 03:35:44 crc kubenswrapper[4980]: E1206 03:35:44.184323 4980 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-mcprh" podUID="e6e88c5e-a255-449a-ae38-ac4d73a8e984" Dec 06 03:35:44 crc kubenswrapper[4980]: E1206 03:35:44.184437 4980 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 03:35:45 crc kubenswrapper[4980]: I1206 03:35:45.183729 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 03:35:45 crc kubenswrapper[4980]: E1206 03:35:45.184219 4980 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 03:35:46 crc kubenswrapper[4980]: I1206 03:35:46.184373 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 03:35:46 crc kubenswrapper[4980]: I1206 03:35:46.184418 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 03:35:46 crc kubenswrapper[4980]: I1206 03:35:46.184423 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-mcprh" Dec 06 03:35:46 crc kubenswrapper[4980]: E1206 03:35:46.184569 4980 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 03:35:46 crc kubenswrapper[4980]: E1206 03:35:46.185078 4980 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 03:35:46 crc kubenswrapper[4980]: E1206 03:35:46.185189 4980 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-mcprh" podUID="e6e88c5e-a255-449a-ae38-ac4d73a8e984" Dec 06 03:35:46 crc kubenswrapper[4980]: I1206 03:35:46.185780 4980 scope.go:117] "RemoveContainer" containerID="33fa9bd73b9af5df6de4e7ed019a512334f465b830a3f4396cde14989658bf55" Dec 06 03:35:46 crc kubenswrapper[4980]: E1206 03:35:46.186064 4980 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-5w4fr_openshift-ovn-kubernetes(5af0e768-3c9c-4401-ab56-57516cd3170e)\"" pod="openshift-ovn-kubernetes/ovnkube-node-5w4fr" podUID="5af0e768-3c9c-4401-ab56-57516cd3170e" Dec 06 03:35:47 crc kubenswrapper[4980]: I1206 03:35:47.183980 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 03:35:47 crc kubenswrapper[4980]: E1206 03:35:47.184477 4980 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 03:35:48 crc kubenswrapper[4980]: I1206 03:35:48.183622 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 03:35:48 crc kubenswrapper[4980]: I1206 03:35:48.183688 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 03:35:48 crc kubenswrapper[4980]: I1206 03:35:48.183718 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-mcprh" Dec 06 03:35:48 crc kubenswrapper[4980]: E1206 03:35:48.184163 4980 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 03:35:48 crc kubenswrapper[4980]: E1206 03:35:48.184245 4980 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-mcprh" podUID="e6e88c5e-a255-449a-ae38-ac4d73a8e984" Dec 06 03:35:48 crc kubenswrapper[4980]: E1206 03:35:48.184582 4980 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 03:35:49 crc kubenswrapper[4980]: I1206 03:35:49.183634 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 03:35:49 crc kubenswrapper[4980]: E1206 03:35:49.183766 4980 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 03:35:50 crc kubenswrapper[4980]: I1206 03:35:50.183773 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-mcprh" Dec 06 03:35:50 crc kubenswrapper[4980]: E1206 03:35:50.183913 4980 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-mcprh" podUID="e6e88c5e-a255-449a-ae38-ac4d73a8e984" Dec 06 03:35:50 crc kubenswrapper[4980]: I1206 03:35:50.184237 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 03:35:50 crc kubenswrapper[4980]: E1206 03:35:50.184297 4980 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 03:35:50 crc kubenswrapper[4980]: I1206 03:35:50.184685 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 03:35:50 crc kubenswrapper[4980]: E1206 03:35:50.184780 4980 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 03:35:51 crc kubenswrapper[4980]: I1206 03:35:51.184038 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 03:35:51 crc kubenswrapper[4980]: E1206 03:35:51.186074 4980 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 03:35:52 crc kubenswrapper[4980]: I1206 03:35:52.184394 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 03:35:52 crc kubenswrapper[4980]: I1206 03:35:52.184940 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 03:35:52 crc kubenswrapper[4980]: I1206 03:35:52.184406 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-mcprh" Dec 06 03:35:52 crc kubenswrapper[4980]: E1206 03:35:52.185054 4980 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 03:35:52 crc kubenswrapper[4980]: E1206 03:35:52.185132 4980 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-mcprh" podUID="e6e88c5e-a255-449a-ae38-ac4d73a8e984" Dec 06 03:35:52 crc kubenswrapper[4980]: E1206 03:35:52.185272 4980 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 03:35:53 crc kubenswrapper[4980]: I1206 03:35:53.184136 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 03:35:53 crc kubenswrapper[4980]: E1206 03:35:53.184284 4980 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 03:35:54 crc kubenswrapper[4980]: I1206 03:35:54.183793 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 03:35:54 crc kubenswrapper[4980]: I1206 03:35:54.183806 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 03:35:54 crc kubenswrapper[4980]: E1206 03:35:54.183981 4980 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 03:35:54 crc kubenswrapper[4980]: I1206 03:35:54.183810 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-mcprh" Dec 06 03:35:54 crc kubenswrapper[4980]: E1206 03:35:54.184135 4980 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 03:35:54 crc kubenswrapper[4980]: E1206 03:35:54.184261 4980 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-mcprh" podUID="e6e88c5e-a255-449a-ae38-ac4d73a8e984" Dec 06 03:35:55 crc kubenswrapper[4980]: I1206 03:35:55.184320 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 03:35:55 crc kubenswrapper[4980]: E1206 03:35:55.184475 4980 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 03:35:56 crc kubenswrapper[4980]: I1206 03:35:56.183893 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-mcprh" Dec 06 03:35:56 crc kubenswrapper[4980]: I1206 03:35:56.184032 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 03:35:56 crc kubenswrapper[4980]: E1206 03:35:56.184126 4980 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-mcprh" podUID="e6e88c5e-a255-449a-ae38-ac4d73a8e984" Dec 06 03:35:56 crc kubenswrapper[4980]: I1206 03:35:56.184141 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 03:35:56 crc kubenswrapper[4980]: E1206 03:35:56.184482 4980 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 03:35:56 crc kubenswrapper[4980]: E1206 03:35:56.184643 4980 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 03:35:57 crc kubenswrapper[4980]: I1206 03:35:57.183804 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 03:35:57 crc kubenswrapper[4980]: E1206 03:35:57.184381 4980 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 03:35:58 crc kubenswrapper[4980]: I1206 03:35:58.184053 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 03:35:58 crc kubenswrapper[4980]: I1206 03:35:58.184053 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 03:35:58 crc kubenswrapper[4980]: I1206 03:35:58.184078 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-mcprh" Dec 06 03:35:58 crc kubenswrapper[4980]: E1206 03:35:58.184271 4980 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 03:35:58 crc kubenswrapper[4980]: E1206 03:35:58.184346 4980 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-mcprh" podUID="e6e88c5e-a255-449a-ae38-ac4d73a8e984" Dec 06 03:35:58 crc kubenswrapper[4980]: E1206 03:35:58.184454 4980 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 03:35:59 crc kubenswrapper[4980]: I1206 03:35:59.183824 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 03:35:59 crc kubenswrapper[4980]: E1206 03:35:59.184357 4980 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 03:36:00 crc kubenswrapper[4980]: I1206 03:36:00.183797 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 03:36:00 crc kubenswrapper[4980]: I1206 03:36:00.183827 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-mcprh" Dec 06 03:36:00 crc kubenswrapper[4980]: E1206 03:36:00.183969 4980 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 03:36:00 crc kubenswrapper[4980]: I1206 03:36:00.183809 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 03:36:00 crc kubenswrapper[4980]: E1206 03:36:00.184050 4980 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-mcprh" podUID="e6e88c5e-a255-449a-ae38-ac4d73a8e984" Dec 06 03:36:00 crc kubenswrapper[4980]: E1206 03:36:00.184236 4980 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 03:36:01 crc kubenswrapper[4980]: I1206 03:36:01.184118 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 03:36:01 crc kubenswrapper[4980]: E1206 03:36:01.185791 4980 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 03:36:01 crc kubenswrapper[4980]: I1206 03:36:01.186060 4980 scope.go:117] "RemoveContainer" containerID="33fa9bd73b9af5df6de4e7ed019a512334f465b830a3f4396cde14989658bf55" Dec 06 03:36:01 crc kubenswrapper[4980]: E1206 03:36:01.186260 4980 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-5w4fr_openshift-ovn-kubernetes(5af0e768-3c9c-4401-ab56-57516cd3170e)\"" pod="openshift-ovn-kubernetes/ovnkube-node-5w4fr" podUID="5af0e768-3c9c-4401-ab56-57516cd3170e" Dec 06 03:36:01 crc kubenswrapper[4980]: E1206 03:36:01.193769 4980 kubelet_node_status.go:497] "Node not becoming ready in time after startup" Dec 06 03:36:01 crc kubenswrapper[4980]: E1206 03:36:01.492656 4980 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 06 03:36:02 crc kubenswrapper[4980]: I1206 03:36:02.183670 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 03:36:02 crc kubenswrapper[4980]: I1206 03:36:02.183680 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 03:36:02 crc kubenswrapper[4980]: I1206 03:36:02.183825 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-mcprh" Dec 06 03:36:02 crc kubenswrapper[4980]: E1206 03:36:02.184015 4980 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 03:36:02 crc kubenswrapper[4980]: E1206 03:36:02.184208 4980 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 03:36:02 crc kubenswrapper[4980]: E1206 03:36:02.184328 4980 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-mcprh" podUID="e6e88c5e-a255-449a-ae38-ac4d73a8e984" Dec 06 03:36:02 crc kubenswrapper[4980]: I1206 03:36:02.315337 4980 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-kq4z8_f96c1f6c-55e8-4d86-bff3-5381581b7022/kube-multus/1.log" Dec 06 03:36:02 crc kubenswrapper[4980]: I1206 03:36:02.316248 4980 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-kq4z8_f96c1f6c-55e8-4d86-bff3-5381581b7022/kube-multus/0.log" Dec 06 03:36:02 crc kubenswrapper[4980]: I1206 03:36:02.316409 4980 generic.go:334] "Generic (PLEG): container finished" podID="f96c1f6c-55e8-4d86-bff3-5381581b7022" containerID="ea7f0bc23247b5effe7f0250582c8694bbf0e9a2d185a844773665a0a48f350e" exitCode=1 Dec 06 03:36:02 crc kubenswrapper[4980]: I1206 03:36:02.316575 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-kq4z8" event={"ID":"f96c1f6c-55e8-4d86-bff3-5381581b7022","Type":"ContainerDied","Data":"ea7f0bc23247b5effe7f0250582c8694bbf0e9a2d185a844773665a0a48f350e"} Dec 06 03:36:02 crc kubenswrapper[4980]: I1206 03:36:02.316685 4980 scope.go:117] "RemoveContainer" containerID="799498681afa7653408bc23b1e82d75c16e3c809d2ee7f322780e20c537eb710" Dec 06 03:36:02 crc kubenswrapper[4980]: I1206 03:36:02.317346 4980 scope.go:117] "RemoveContainer" containerID="ea7f0bc23247b5effe7f0250582c8694bbf0e9a2d185a844773665a0a48f350e" Dec 06 03:36:02 crc kubenswrapper[4980]: E1206 03:36:02.317717 4980 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-multus pod=multus-kq4z8_openshift-multus(f96c1f6c-55e8-4d86-bff3-5381581b7022)\"" pod="openshift-multus/multus-kq4z8" podUID="f96c1f6c-55e8-4d86-bff3-5381581b7022" Dec 06 03:36:03 crc kubenswrapper[4980]: I1206 03:36:03.184076 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 03:36:03 crc kubenswrapper[4980]: E1206 03:36:03.184276 4980 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 03:36:03 crc kubenswrapper[4980]: I1206 03:36:03.323862 4980 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-kq4z8_f96c1f6c-55e8-4d86-bff3-5381581b7022/kube-multus/1.log" Dec 06 03:36:04 crc kubenswrapper[4980]: I1206 03:36:04.183811 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 03:36:04 crc kubenswrapper[4980]: I1206 03:36:04.183821 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 03:36:04 crc kubenswrapper[4980]: I1206 03:36:04.183842 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-mcprh" Dec 06 03:36:04 crc kubenswrapper[4980]: E1206 03:36:04.183998 4980 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 03:36:04 crc kubenswrapper[4980]: E1206 03:36:04.184237 4980 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 03:36:04 crc kubenswrapper[4980]: E1206 03:36:04.184326 4980 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-mcprh" podUID="e6e88c5e-a255-449a-ae38-ac4d73a8e984" Dec 06 03:36:05 crc kubenswrapper[4980]: I1206 03:36:05.184419 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 03:36:05 crc kubenswrapper[4980]: E1206 03:36:05.184652 4980 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 03:36:06 crc kubenswrapper[4980]: I1206 03:36:06.184012 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 03:36:06 crc kubenswrapper[4980]: I1206 03:36:06.184060 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 03:36:06 crc kubenswrapper[4980]: I1206 03:36:06.184134 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-mcprh" Dec 06 03:36:06 crc kubenswrapper[4980]: E1206 03:36:06.184153 4980 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 03:36:06 crc kubenswrapper[4980]: E1206 03:36:06.184191 4980 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 03:36:06 crc kubenswrapper[4980]: E1206 03:36:06.184422 4980 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-mcprh" podUID="e6e88c5e-a255-449a-ae38-ac4d73a8e984" Dec 06 03:36:06 crc kubenswrapper[4980]: E1206 03:36:06.494384 4980 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 06 03:36:07 crc kubenswrapper[4980]: I1206 03:36:07.183626 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 03:36:07 crc kubenswrapper[4980]: E1206 03:36:07.183841 4980 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 03:36:08 crc kubenswrapper[4980]: I1206 03:36:08.184273 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 03:36:08 crc kubenswrapper[4980]: I1206 03:36:08.184405 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-mcprh" Dec 06 03:36:08 crc kubenswrapper[4980]: E1206 03:36:08.184539 4980 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 03:36:08 crc kubenswrapper[4980]: I1206 03:36:08.184399 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 03:36:08 crc kubenswrapper[4980]: E1206 03:36:08.184725 4980 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-mcprh" podUID="e6e88c5e-a255-449a-ae38-ac4d73a8e984" Dec 06 03:36:08 crc kubenswrapper[4980]: E1206 03:36:08.184989 4980 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 03:36:09 crc kubenswrapper[4980]: I1206 03:36:09.184306 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 03:36:09 crc kubenswrapper[4980]: E1206 03:36:09.184541 4980 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 03:36:10 crc kubenswrapper[4980]: I1206 03:36:10.183855 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 03:36:10 crc kubenswrapper[4980]: I1206 03:36:10.183883 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 03:36:10 crc kubenswrapper[4980]: I1206 03:36:10.183947 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-mcprh" Dec 06 03:36:10 crc kubenswrapper[4980]: E1206 03:36:10.184182 4980 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 03:36:10 crc kubenswrapper[4980]: E1206 03:36:10.184318 4980 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 03:36:10 crc kubenswrapper[4980]: E1206 03:36:10.184505 4980 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-mcprh" podUID="e6e88c5e-a255-449a-ae38-ac4d73a8e984" Dec 06 03:36:11 crc kubenswrapper[4980]: I1206 03:36:11.184251 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 03:36:11 crc kubenswrapper[4980]: E1206 03:36:11.185573 4980 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 03:36:11 crc kubenswrapper[4980]: E1206 03:36:11.495454 4980 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 06 03:36:12 crc kubenswrapper[4980]: I1206 03:36:12.183686 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 03:36:12 crc kubenswrapper[4980]: I1206 03:36:12.183686 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 03:36:12 crc kubenswrapper[4980]: I1206 03:36:12.184429 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-mcprh" Dec 06 03:36:12 crc kubenswrapper[4980]: E1206 03:36:12.184729 4980 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 03:36:12 crc kubenswrapper[4980]: E1206 03:36:12.184894 4980 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 03:36:12 crc kubenswrapper[4980]: E1206 03:36:12.185104 4980 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-mcprh" podUID="e6e88c5e-a255-449a-ae38-ac4d73a8e984" Dec 06 03:36:12 crc kubenswrapper[4980]: I1206 03:36:12.186965 4980 scope.go:117] "RemoveContainer" containerID="33fa9bd73b9af5df6de4e7ed019a512334f465b830a3f4396cde14989658bf55" Dec 06 03:36:12 crc kubenswrapper[4980]: I1206 03:36:12.360356 4980 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-5w4fr_5af0e768-3c9c-4401-ab56-57516cd3170e/ovnkube-controller/3.log" Dec 06 03:36:13 crc kubenswrapper[4980]: I1206 03:36:13.186410 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 03:36:13 crc kubenswrapper[4980]: E1206 03:36:13.186557 4980 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 03:36:13 crc kubenswrapper[4980]: I1206 03:36:13.338508 4980 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-mcprh"] Dec 06 03:36:13 crc kubenswrapper[4980]: I1206 03:36:13.338692 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-mcprh" Dec 06 03:36:13 crc kubenswrapper[4980]: E1206 03:36:13.338836 4980 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-mcprh" podUID="e6e88c5e-a255-449a-ae38-ac4d73a8e984" Dec 06 03:36:13 crc kubenswrapper[4980]: I1206 03:36:13.371719 4980 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-5w4fr_5af0e768-3c9c-4401-ab56-57516cd3170e/ovnkube-controller/3.log" Dec 06 03:36:13 crc kubenswrapper[4980]: I1206 03:36:13.374576 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5w4fr" event={"ID":"5af0e768-3c9c-4401-ab56-57516cd3170e","Type":"ContainerStarted","Data":"2d3ca7fa81b93eca00b5d90c8b2685ff607a8156fe318cf99c5296fd65a5be75"} Dec 06 03:36:13 crc kubenswrapper[4980]: I1206 03:36:13.375103 4980 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-5w4fr" Dec 06 03:36:13 crc kubenswrapper[4980]: I1206 03:36:13.416850 4980 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-5w4fr" podStartSLOduration=111.416822506 podStartE2EDuration="1m51.416822506s" podCreationTimestamp="2025-12-06 03:34:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 03:36:13.416219389 +0000 UTC m=+132.660970690" watchObservedRunningTime="2025-12-06 03:36:13.416822506 +0000 UTC m=+132.661573787" Dec 06 03:36:14 crc kubenswrapper[4980]: I1206 03:36:14.183761 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 03:36:14 crc kubenswrapper[4980]: I1206 03:36:14.183760 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 03:36:14 crc kubenswrapper[4980]: E1206 03:36:14.184390 4980 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 03:36:14 crc kubenswrapper[4980]: E1206 03:36:14.184548 4980 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 03:36:15 crc kubenswrapper[4980]: I1206 03:36:15.183726 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-mcprh" Dec 06 03:36:15 crc kubenswrapper[4980]: I1206 03:36:15.183755 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 03:36:15 crc kubenswrapper[4980]: E1206 03:36:15.183877 4980 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-mcprh" podUID="e6e88c5e-a255-449a-ae38-ac4d73a8e984" Dec 06 03:36:15 crc kubenswrapper[4980]: E1206 03:36:15.184170 4980 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 03:36:15 crc kubenswrapper[4980]: I1206 03:36:15.184810 4980 scope.go:117] "RemoveContainer" containerID="ea7f0bc23247b5effe7f0250582c8694bbf0e9a2d185a844773665a0a48f350e" Dec 06 03:36:16 crc kubenswrapper[4980]: I1206 03:36:16.183953 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 03:36:16 crc kubenswrapper[4980]: I1206 03:36:16.183969 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 03:36:16 crc kubenswrapper[4980]: E1206 03:36:16.184150 4980 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 03:36:16 crc kubenswrapper[4980]: E1206 03:36:16.184301 4980 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 03:36:16 crc kubenswrapper[4980]: I1206 03:36:16.389743 4980 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-kq4z8_f96c1f6c-55e8-4d86-bff3-5381581b7022/kube-multus/1.log" Dec 06 03:36:16 crc kubenswrapper[4980]: I1206 03:36:16.389828 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-kq4z8" event={"ID":"f96c1f6c-55e8-4d86-bff3-5381581b7022","Type":"ContainerStarted","Data":"f894aadbd8dc641edcede1ab30092ef8087c9314de751da017e522e14405996a"} Dec 06 03:36:16 crc kubenswrapper[4980]: E1206 03:36:16.496444 4980 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 06 03:36:17 crc kubenswrapper[4980]: I1206 03:36:17.183609 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-mcprh" Dec 06 03:36:17 crc kubenswrapper[4980]: I1206 03:36:17.183664 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 03:36:17 crc kubenswrapper[4980]: E1206 03:36:17.183786 4980 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-mcprh" podUID="e6e88c5e-a255-449a-ae38-ac4d73a8e984" Dec 06 03:36:17 crc kubenswrapper[4980]: E1206 03:36:17.183932 4980 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 03:36:18 crc kubenswrapper[4980]: I1206 03:36:18.183957 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 03:36:18 crc kubenswrapper[4980]: I1206 03:36:18.184058 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 03:36:18 crc kubenswrapper[4980]: E1206 03:36:18.184171 4980 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 03:36:18 crc kubenswrapper[4980]: E1206 03:36:18.184375 4980 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 03:36:19 crc kubenswrapper[4980]: I1206 03:36:19.183899 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 03:36:19 crc kubenswrapper[4980]: I1206 03:36:19.184004 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-mcprh" Dec 06 03:36:19 crc kubenswrapper[4980]: E1206 03:36:19.184057 4980 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 03:36:19 crc kubenswrapper[4980]: E1206 03:36:19.184199 4980 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-mcprh" podUID="e6e88c5e-a255-449a-ae38-ac4d73a8e984" Dec 06 03:36:20 crc kubenswrapper[4980]: I1206 03:36:20.183860 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 03:36:20 crc kubenswrapper[4980]: I1206 03:36:20.183860 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 03:36:20 crc kubenswrapper[4980]: E1206 03:36:20.184086 4980 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 03:36:20 crc kubenswrapper[4980]: E1206 03:36:20.184232 4980 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 03:36:21 crc kubenswrapper[4980]: I1206 03:36:21.184218 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-mcprh" Dec 06 03:36:21 crc kubenswrapper[4980]: I1206 03:36:21.184234 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 03:36:21 crc kubenswrapper[4980]: E1206 03:36:21.186600 4980 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-mcprh" podUID="e6e88c5e-a255-449a-ae38-ac4d73a8e984" Dec 06 03:36:21 crc kubenswrapper[4980]: E1206 03:36:21.187029 4980 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 03:36:22 crc kubenswrapper[4980]: I1206 03:36:22.184437 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 03:36:22 crc kubenswrapper[4980]: I1206 03:36:22.184814 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 03:36:22 crc kubenswrapper[4980]: I1206 03:36:22.187598 4980 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Dec 06 03:36:22 crc kubenswrapper[4980]: I1206 03:36:22.187633 4980 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Dec 06 03:36:22 crc kubenswrapper[4980]: I1206 03:36:22.188357 4980 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Dec 06 03:36:22 crc kubenswrapper[4980]: I1206 03:36:22.190028 4980 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Dec 06 03:36:23 crc kubenswrapper[4980]: I1206 03:36:23.184090 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 03:36:23 crc kubenswrapper[4980]: I1206 03:36:23.184655 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-mcprh" Dec 06 03:36:23 crc kubenswrapper[4980]: I1206 03:36:23.187910 4980 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Dec 06 03:36:23 crc kubenswrapper[4980]: I1206 03:36:23.188441 4980 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Dec 06 03:36:25 crc kubenswrapper[4980]: I1206 03:36:25.137686 4980 patch_prober.go:28] interesting pod/machine-config-daemon-r5zfb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 03:36:25 crc kubenswrapper[4980]: I1206 03:36:25.137805 4980 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-r5zfb" podUID="320f44d1-a671-4a91-b328-a8b0fdd8f23a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 03:36:30 crc kubenswrapper[4980]: I1206 03:36:30.190803 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 03:36:30 crc kubenswrapper[4980]: I1206 03:36:30.191058 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 03:36:30 crc kubenswrapper[4980]: I1206 03:36:30.192373 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 03:36:30 crc kubenswrapper[4980]: I1206 03:36:30.199096 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 03:36:30 crc kubenswrapper[4980]: I1206 03:36:30.311373 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 03:36:30 crc kubenswrapper[4980]: I1206 03:36:30.394322 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 03:36:30 crc kubenswrapper[4980]: E1206 03:36:30.394755 4980 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 03:38:32.394704923 +0000 UTC m=+271.639456234 (durationBeforeRetry 2m2s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 03:36:30 crc kubenswrapper[4980]: I1206 03:36:30.395018 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 03:36:30 crc kubenswrapper[4980]: I1206 03:36:30.395098 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 03:36:30 crc kubenswrapper[4980]: I1206 03:36:30.401012 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 03:36:30 crc kubenswrapper[4980]: I1206 03:36:30.402334 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 03:36:30 crc kubenswrapper[4980]: I1206 03:36:30.404677 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 03:36:30 crc kubenswrapper[4980]: I1206 03:36:30.626621 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 03:36:30 crc kubenswrapper[4980]: W1206 03:36:30.630629 4980 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5fe485a1_e14f_4c09_b5b9_f252bc42b7e8.slice/crio-5d7eb89c42cd753080349aef30db6565512f9b5099c4595b53596f5cc3e52ae9 WatchSource:0}: Error finding container 5d7eb89c42cd753080349aef30db6565512f9b5099c4595b53596f5cc3e52ae9: Status 404 returned error can't find the container with id 5d7eb89c42cd753080349aef30db6565512f9b5099c4595b53596f5cc3e52ae9 Dec 06 03:36:30 crc kubenswrapper[4980]: W1206 03:36:30.655726 4980 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3b6479f0_333b_4a96_9adf_2099afdc2447.slice/crio-452c441f3f24db96ee4c7585a33441e4206c9b9a2d573710d6f584e969c466dd WatchSource:0}: Error finding container 452c441f3f24db96ee4c7585a33441e4206c9b9a2d573710d6f584e969c466dd: Status 404 returned error can't find the container with id 452c441f3f24db96ee4c7585a33441e4206c9b9a2d573710d6f584e969c466dd Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.221620 4980 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeReady" Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.262234 4980 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-s6h9g"] Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.262903 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-s6h9g" Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.265119 4980 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-qpchc"] Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.265486 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-qpchc" Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.278550 4980 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.278992 4980 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Dec 06 03:36:31 crc kubenswrapper[4980]: W1206 03:36:31.279206 4980 reflector.go:561] object-"openshift-apiserver"/"image-import-ca": failed to list *v1.ConfigMap: configmaps "image-import-ca" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-apiserver": no relationship found between node 'crc' and this object Dec 06 03:36:31 crc kubenswrapper[4980]: E1206 03:36:31.279245 4980 reflector.go:158] "Unhandled Error" err="object-\"openshift-apiserver\"/\"image-import-ca\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"image-import-ca\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-apiserver\": no relationship found between node 'crc' and this object" logger="UnhandledError" Dec 06 03:36:31 crc kubenswrapper[4980]: W1206 03:36:31.289753 4980 reflector.go:561] object-"openshift-apiserver"/"etcd-client": failed to list *v1.Secret: secrets "etcd-client" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "openshift-apiserver": no relationship found between node 'crc' and this object Dec 06 03:36:31 crc kubenswrapper[4980]: E1206 03:36:31.289826 4980 reflector.go:158] "Unhandled Error" err="object-\"openshift-apiserver\"/\"etcd-client\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"etcd-client\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"openshift-apiserver\": no relationship found between node 'crc' and this object" logger="UnhandledError" Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.290052 4980 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.290219 4980 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-f9d7485db-gjmj8"] Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.290799 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-gjmj8" Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.290211 4980 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.292698 4980 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-machine-approver/machine-approver-56656f9798-425jl"] Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.293754 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-425jl" Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.293918 4980 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-5zm4c"] Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.294543 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-5zm4c" Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.295346 4980 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.295494 4980 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.296023 4980 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-2ph5v"] Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.296447 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-2ph5v" Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.297704 4980 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.297994 4980 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.298204 4980 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.298373 4980 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.298535 4980 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.301756 4980 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/downloads-7954f5f757-6h96g"] Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.302137 4980 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-84ft2"] Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.302376 4980 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-x5m4p"] Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.302790 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-x5m4p" Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.302886 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-6h96g" Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.303445 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-84ft2" Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.303916 4980 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.304217 4980 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.304335 4980 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-9fg2s"] Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.304450 4980 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.304987 4980 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-zqsvr"] Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.305596 4980 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-pmmcw"] Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.306105 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-pmmcw" Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.306591 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-9fg2s" Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.307339 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-zqsvr" Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.307404 4980 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-66bs5"] Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.308064 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/2cf31abb-ef18-4ba5-8cc9-bf3f7507f771-trusted-ca-bundle\") pod \"apiserver-76f77b778f-s6h9g\" (UID: \"2cf31abb-ef18-4ba5-8cc9-bf3f7507f771\") " pod="openshift-apiserver/apiserver-76f77b778f-s6h9g" Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.308100 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/2cf31abb-ef18-4ba5-8cc9-bf3f7507f771-audit\") pod \"apiserver-76f77b778f-s6h9g\" (UID: \"2cf31abb-ef18-4ba5-8cc9-bf3f7507f771\") " pod="openshift-apiserver/apiserver-76f77b778f-s6h9g" Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.308126 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/2cf31abb-ef18-4ba5-8cc9-bf3f7507f771-node-pullsecrets\") pod \"apiserver-76f77b778f-s6h9g\" (UID: \"2cf31abb-ef18-4ba5-8cc9-bf3f7507f771\") " pod="openshift-apiserver/apiserver-76f77b778f-s6h9g" Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.308149 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/2cf31abb-ef18-4ba5-8cc9-bf3f7507f771-image-import-ca\") pod \"apiserver-76f77b778f-s6h9g\" (UID: \"2cf31abb-ef18-4ba5-8cc9-bf3f7507f771\") " pod="openshift-apiserver/apiserver-76f77b778f-s6h9g" Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.308167 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/2cf31abb-ef18-4ba5-8cc9-bf3f7507f771-audit-dir\") pod \"apiserver-76f77b778f-s6h9g\" (UID: \"2cf31abb-ef18-4ba5-8cc9-bf3f7507f771\") " pod="openshift-apiserver/apiserver-76f77b778f-s6h9g" Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.308191 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-66bs5" Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.308191 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/2cf31abb-ef18-4ba5-8cc9-bf3f7507f771-etcd-client\") pod \"apiserver-76f77b778f-s6h9g\" (UID: \"2cf31abb-ef18-4ba5-8cc9-bf3f7507f771\") " pod="openshift-apiserver/apiserver-76f77b778f-s6h9g" Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.308600 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/2cf31abb-ef18-4ba5-8cc9-bf3f7507f771-encryption-config\") pod \"apiserver-76f77b778f-s6h9g\" (UID: \"2cf31abb-ef18-4ba5-8cc9-bf3f7507f771\") " pod="openshift-apiserver/apiserver-76f77b778f-s6h9g" Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.308647 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6l7xr\" (UniqueName: \"kubernetes.io/projected/2cf31abb-ef18-4ba5-8cc9-bf3f7507f771-kube-api-access-6l7xr\") pod \"apiserver-76f77b778f-s6h9g\" (UID: \"2cf31abb-ef18-4ba5-8cc9-bf3f7507f771\") " pod="openshift-apiserver/apiserver-76f77b778f-s6h9g" Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.308654 4980 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-6txpx"] Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.308666 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2cf31abb-ef18-4ba5-8cc9-bf3f7507f771-config\") pod \"apiserver-76f77b778f-s6h9g\" (UID: \"2cf31abb-ef18-4ba5-8cc9-bf3f7507f771\") " pod="openshift-apiserver/apiserver-76f77b778f-s6h9g" Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.308713 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/2cf31abb-ef18-4ba5-8cc9-bf3f7507f771-etcd-serving-ca\") pod \"apiserver-76f77b778f-s6h9g\" (UID: \"2cf31abb-ef18-4ba5-8cc9-bf3f7507f771\") " pod="openshift-apiserver/apiserver-76f77b778f-s6h9g" Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.308741 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2cf31abb-ef18-4ba5-8cc9-bf3f7507f771-serving-cert\") pod \"apiserver-76f77b778f-s6h9g\" (UID: \"2cf31abb-ef18-4ba5-8cc9-bf3f7507f771\") " pod="openshift-apiserver/apiserver-76f77b778f-s6h9g" Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.308992 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-6txpx" Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.309181 4980 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-qw2fb"] Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.309550 4980 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.310788 4980 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.311040 4980 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.311127 4980 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.311252 4980 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.311363 4980 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.311534 4980 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.311671 4980 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.311792 4980 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.311886 4980 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.311957 4980 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.311956 4980 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.312115 4980 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.312158 4980 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.312222 4980 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.312293 4980 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.312328 4980 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.312450 4980 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.312598 4980 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.312711 4980 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.312895 4980 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.314238 4980 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.314348 4980 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.314645 4980 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.314770 4980 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.314870 4980 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.314966 4980 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.314982 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-qw2fb" Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.315114 4980 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.315217 4980 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.319250 4980 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-w6bmg"] Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.321399 4980 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-lrnhm"] Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.322162 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-lrnhm" Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.322475 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-w6bmg" Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.325032 4980 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.325033 4980 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.325364 4980 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.325292 4980 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.326489 4980 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console-operator/console-operator-58897d9998-kftgf"] Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.334634 4980 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.336556 4980 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.340007 4980 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.345809 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-kftgf" Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.402953 4980 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.403352 4980 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.403481 4980 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-p4gmt"] Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.404628 4980 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.404743 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-p4gmt" Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.404812 4980 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.404648 4980 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-xjtnv"] Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.404656 4980 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.406298 4980 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress/router-default-5444994796-pzzkd"] Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.406840 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-pzzkd" Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.407118 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-xjtnv" Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.407486 4980 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.407649 4980 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.409721 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/2cf31abb-ef18-4ba5-8cc9-bf3f7507f771-audit-dir\") pod \"apiserver-76f77b778f-s6h9g\" (UID: \"2cf31abb-ef18-4ba5-8cc9-bf3f7507f771\") " pod="openshift-apiserver/apiserver-76f77b778f-s6h9g" Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.409754 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/2ac9a45e-ea4a-472e-bd5a-d947a0e2713e-client-ca\") pod \"controller-manager-879f6c89f-qpchc\" (UID: \"2ac9a45e-ea4a-472e-bd5a-d947a0e2713e\") " pod="openshift-controller-manager/controller-manager-879f6c89f-qpchc" Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.409774 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/005c8e70-2c2b-43a4-ae31-fd9013840824-auth-proxy-config\") pod \"machine-approver-56656f9798-425jl\" (UID: \"005c8e70-2c2b-43a4-ae31-fd9013840824\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-425jl" Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.409793 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f3921e9a-9d9a-45e0-8733-6a1cd20422d2-serving-cert\") pod \"etcd-operator-b45778765-p4gmt\" (UID: \"f3921e9a-9d9a-45e0-8733-6a1cd20422d2\") " pod="openshift-etcd-operator/etcd-operator-b45778765-p4gmt" Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.409809 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/2cf31abb-ef18-4ba5-8cc9-bf3f7507f771-encryption-config\") pod \"apiserver-76f77b778f-s6h9g\" (UID: \"2cf31abb-ef18-4ba5-8cc9-bf3f7507f771\") " pod="openshift-apiserver/apiserver-76f77b778f-s6h9g" Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.409827 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/443a2dc9-0546-4ab0-a448-eab81c09b482-console-config\") pod \"console-f9d7485db-gjmj8\" (UID: \"443a2dc9-0546-4ab0-a448-eab81c09b482\") " pod="openshift-console/console-f9d7485db-gjmj8" Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.409843 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f2ea7661-b67f-4fbb-924a-943fe48995cb-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-6txpx\" (UID: \"f2ea7661-b67f-4fbb-924a-943fe48995cb\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-6txpx" Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.409859 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f8973f62-160b-4c4b-9238-6a2ff3d63bc5-config\") pod \"machine-api-operator-5694c8668f-x5m4p\" (UID: \"f8973f62-160b-4c4b-9238-6a2ff3d63bc5\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-x5m4p" Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.409865 4980 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.409878 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2ac9a45e-ea4a-472e-bd5a-d947a0e2713e-config\") pod \"controller-manager-879f6c89f-qpchc\" (UID: \"2ac9a45e-ea4a-472e-bd5a-d947a0e2713e\") " pod="openshift-controller-manager/controller-manager-879f6c89f-qpchc" Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.409895 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a83bd6cc-62d8-44f4-930b-1c8fb70277c5-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-9fg2s\" (UID: \"a83bd6cc-62d8-44f4-930b-1c8fb70277c5\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-9fg2s" Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.409912 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/41587982-d7bb-413c-a1b4-12baa2ea600a-encryption-config\") pod \"apiserver-7bbb656c7d-66bs5\" (UID: \"41587982-d7bb-413c-a1b4-12baa2ea600a\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-66bs5" Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.409928 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/41587982-d7bb-413c-a1b4-12baa2ea600a-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-66bs5\" (UID: \"41587982-d7bb-413c-a1b4-12baa2ea600a\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-66bs5" Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.409946 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dqls7\" (UniqueName: \"kubernetes.io/projected/41587982-d7bb-413c-a1b4-12baa2ea600a-kube-api-access-dqls7\") pod \"apiserver-7bbb656c7d-66bs5\" (UID: \"41587982-d7bb-413c-a1b4-12baa2ea600a\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-66bs5" Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.409960 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/443a2dc9-0546-4ab0-a448-eab81c09b482-service-ca\") pod \"console-f9d7485db-gjmj8\" (UID: \"443a2dc9-0546-4ab0-a448-eab81c09b482\") " pod="openshift-console/console-f9d7485db-gjmj8" Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.409977 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c7db3f86-0e71-46d1-b95c-929069bf8b3a-serving-cert\") pod \"route-controller-manager-6576b87f9c-2ph5v\" (UID: \"c7db3f86-0e71-46d1-b95c-929069bf8b3a\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-2ph5v" Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.409992 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/443a2dc9-0546-4ab0-a448-eab81c09b482-trusted-ca-bundle\") pod \"console-f9d7485db-gjmj8\" (UID: \"443a2dc9-0546-4ab0-a448-eab81c09b482\") " pod="openshift-console/console-f9d7485db-gjmj8" Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.410017 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-svpft\" (UniqueName: \"kubernetes.io/projected/443a2dc9-0546-4ab0-a448-eab81c09b482-kube-api-access-svpft\") pod \"console-f9d7485db-gjmj8\" (UID: \"443a2dc9-0546-4ab0-a448-eab81c09b482\") " pod="openshift-console/console-f9d7485db-gjmj8" Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.410033 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2cf31abb-ef18-4ba5-8cc9-bf3f7507f771-serving-cert\") pod \"apiserver-76f77b778f-s6h9g\" (UID: \"2cf31abb-ef18-4ba5-8cc9-bf3f7507f771\") " pod="openshift-apiserver/apiserver-76f77b778f-s6h9g" Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.410047 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/41587982-d7bb-413c-a1b4-12baa2ea600a-serving-cert\") pod \"apiserver-7bbb656c7d-66bs5\" (UID: \"41587982-d7bb-413c-a1b4-12baa2ea600a\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-66bs5" Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.410061 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/41587982-d7bb-413c-a1b4-12baa2ea600a-audit-dir\") pod \"apiserver-7bbb656c7d-66bs5\" (UID: \"41587982-d7bb-413c-a1b4-12baa2ea600a\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-66bs5" Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.410075 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7c3cb795-33ac-4b16-b41e-a7da0de03578-serving-cert\") pod \"openshift-config-operator-7777fb866f-zqsvr\" (UID: \"7c3cb795-33ac-4b16-b41e-a7da0de03578\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-zqsvr" Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.410091 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dfhqz\" (UniqueName: \"kubernetes.io/projected/7c3cb795-33ac-4b16-b41e-a7da0de03578-kube-api-access-dfhqz\") pod \"openshift-config-operator-7777fb866f-zqsvr\" (UID: \"7c3cb795-33ac-4b16-b41e-a7da0de03578\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-zqsvr" Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.410107 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/f3921e9a-9d9a-45e0-8733-6a1cd20422d2-etcd-client\") pod \"etcd-operator-b45778765-p4gmt\" (UID: \"f3921e9a-9d9a-45e0-8733-6a1cd20422d2\") " pod="openshift-etcd-operator/etcd-operator-b45778765-p4gmt" Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.410123 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/a83bd6cc-62d8-44f4-930b-1c8fb70277c5-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-9fg2s\" (UID: \"a83bd6cc-62d8-44f4-930b-1c8fb70277c5\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-9fg2s" Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.410140 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a83bd6cc-62d8-44f4-930b-1c8fb70277c5-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-9fg2s\" (UID: \"a83bd6cc-62d8-44f4-930b-1c8fb70277c5\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-9fg2s" Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.410154 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/005c8e70-2c2b-43a4-ae31-fd9013840824-config\") pod \"machine-approver-56656f9798-425jl\" (UID: \"005c8e70-2c2b-43a4-ae31-fd9013840824\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-425jl" Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.410172 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dbbbf\" (UniqueName: \"kubernetes.io/projected/f8973f62-160b-4c4b-9238-6a2ff3d63bc5-kube-api-access-dbbbf\") pod \"machine-api-operator-5694c8668f-x5m4p\" (UID: \"f8973f62-160b-4c4b-9238-6a2ff3d63bc5\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-x5m4p" Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.410197 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wmtnf\" (UniqueName: \"kubernetes.io/projected/f3921e9a-9d9a-45e0-8733-6a1cd20422d2-kube-api-access-wmtnf\") pod \"etcd-operator-b45778765-p4gmt\" (UID: \"f3921e9a-9d9a-45e0-8733-6a1cd20422d2\") " pod="openshift-etcd-operator/etcd-operator-b45778765-p4gmt" Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.410216 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/41587982-d7bb-413c-a1b4-12baa2ea600a-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-66bs5\" (UID: \"41587982-d7bb-413c-a1b4-12baa2ea600a\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-66bs5" Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.410238 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/03939884-e3c9-47b9-a7cb-bc19e2f09c9c-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-5zm4c\" (UID: \"03939884-e3c9-47b9-a7cb-bc19e2f09c9c\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-5zm4c" Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.410236 4980 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-49749"] Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.410264 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/2cf31abb-ef18-4ba5-8cc9-bf3f7507f771-audit\") pod \"apiserver-76f77b778f-s6h9g\" (UID: \"2cf31abb-ef18-4ba5-8cc9-bf3f7507f771\") " pod="openshift-apiserver/apiserver-76f77b778f-s6h9g" Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.410284 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/2ac9a45e-ea4a-472e-bd5a-d947a0e2713e-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-qpchc\" (UID: \"2ac9a45e-ea4a-472e-bd5a-d947a0e2713e\") " pod="openshift-controller-manager/controller-manager-879f6c89f-qpchc" Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.410305 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c7db3f86-0e71-46d1-b95c-929069bf8b3a-config\") pod \"route-controller-manager-6576b87f9c-2ph5v\" (UID: \"c7db3f86-0e71-46d1-b95c-929069bf8b3a\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-2ph5v" Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.410324 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/2cf31abb-ef18-4ba5-8cc9-bf3f7507f771-node-pullsecrets\") pod \"apiserver-76f77b778f-s6h9g\" (UID: \"2cf31abb-ef18-4ba5-8cc9-bf3f7507f771\") " pod="openshift-apiserver/apiserver-76f77b778f-s6h9g" Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.410340 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/2cf31abb-ef18-4ba5-8cc9-bf3f7507f771-image-import-ca\") pod \"apiserver-76f77b778f-s6h9g\" (UID: \"2cf31abb-ef18-4ba5-8cc9-bf3f7507f771\") " pod="openshift-apiserver/apiserver-76f77b778f-s6h9g" Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.410356 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/005c8e70-2c2b-43a4-ae31-fd9013840824-machine-approver-tls\") pod \"machine-approver-56656f9798-425jl\" (UID: \"005c8e70-2c2b-43a4-ae31-fd9013840824\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-425jl" Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.410375 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/2cf31abb-ef18-4ba5-8cc9-bf3f7507f771-etcd-client\") pod \"apiserver-76f77b778f-s6h9g\" (UID: \"2cf31abb-ef18-4ba5-8cc9-bf3f7507f771\") " pod="openshift-apiserver/apiserver-76f77b778f-s6h9g" Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.410390 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2ac9a45e-ea4a-472e-bd5a-d947a0e2713e-serving-cert\") pod \"controller-manager-879f6c89f-qpchc\" (UID: \"2ac9a45e-ea4a-472e-bd5a-d947a0e2713e\") " pod="openshift-controller-manager/controller-manager-879f6c89f-qpchc" Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.410406 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/f8973f62-160b-4c4b-9238-6a2ff3d63bc5-images\") pod \"machine-api-operator-5694c8668f-x5m4p\" (UID: \"f8973f62-160b-4c4b-9238-6a2ff3d63bc5\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-x5m4p" Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.410424 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r85l6\" (UniqueName: \"kubernetes.io/projected/2ac9a45e-ea4a-472e-bd5a-d947a0e2713e-kube-api-access-r85l6\") pod \"controller-manager-879f6c89f-qpchc\" (UID: \"2ac9a45e-ea4a-472e-bd5a-d947a0e2713e\") " pod="openshift-controller-manager/controller-manager-879f6c89f-qpchc" Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.410442 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/f3921e9a-9d9a-45e0-8733-6a1cd20422d2-etcd-ca\") pod \"etcd-operator-b45778765-p4gmt\" (UID: \"f3921e9a-9d9a-45e0-8733-6a1cd20422d2\") " pod="openshift-etcd-operator/etcd-operator-b45778765-p4gmt" Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.410459 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/443a2dc9-0546-4ab0-a448-eab81c09b482-oauth-serving-cert\") pod \"console-f9d7485db-gjmj8\" (UID: \"443a2dc9-0546-4ab0-a448-eab81c09b482\") " pod="openshift-console/console-f9d7485db-gjmj8" Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.410476 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n7rjm\" (UniqueName: \"kubernetes.io/projected/c7db3f86-0e71-46d1-b95c-929069bf8b3a-kube-api-access-n7rjm\") pod \"route-controller-manager-6576b87f9c-2ph5v\" (UID: \"c7db3f86-0e71-46d1-b95c-929069bf8b3a\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-2ph5v" Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.410494 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kdpcs\" (UniqueName: \"kubernetes.io/projected/005c8e70-2c2b-43a4-ae31-fd9013840824-kube-api-access-kdpcs\") pod \"machine-approver-56656f9798-425jl\" (UID: \"005c8e70-2c2b-43a4-ae31-fd9013840824\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-425jl" Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.410530 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/7c3cb795-33ac-4b16-b41e-a7da0de03578-available-featuregates\") pod \"openshift-config-operator-7777fb866f-zqsvr\" (UID: \"7c3cb795-33ac-4b16-b41e-a7da0de03578\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-zqsvr" Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.410548 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ndb48\" (UniqueName: \"kubernetes.io/projected/f2ea7661-b67f-4fbb-924a-943fe48995cb-kube-api-access-ndb48\") pod \"openshift-apiserver-operator-796bbdcf4f-6txpx\" (UID: \"f2ea7661-b67f-4fbb-924a-943fe48995cb\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-6txpx" Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.410563 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/f8973f62-160b-4c4b-9238-6a2ff3d63bc5-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-x5m4p\" (UID: \"f8973f62-160b-4c4b-9238-6a2ff3d63bc5\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-x5m4p" Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.410579 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6l7xr\" (UniqueName: \"kubernetes.io/projected/2cf31abb-ef18-4ba5-8cc9-bf3f7507f771-kube-api-access-6l7xr\") pod \"apiserver-76f77b778f-s6h9g\" (UID: \"2cf31abb-ef18-4ba5-8cc9-bf3f7507f771\") " pod="openshift-apiserver/apiserver-76f77b778f-s6h9g" Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.410597 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/443a2dc9-0546-4ab0-a448-eab81c09b482-console-serving-cert\") pod \"console-f9d7485db-gjmj8\" (UID: \"443a2dc9-0546-4ab0-a448-eab81c09b482\") " pod="openshift-console/console-f9d7485db-gjmj8" Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.410612 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/443a2dc9-0546-4ab0-a448-eab81c09b482-console-oauth-config\") pod \"console-f9d7485db-gjmj8\" (UID: \"443a2dc9-0546-4ab0-a448-eab81c09b482\") " pod="openshift-console/console-f9d7485db-gjmj8" Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.410629 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f2ea7661-b67f-4fbb-924a-943fe48995cb-config\") pod \"openshift-apiserver-operator-796bbdcf4f-6txpx\" (UID: \"f2ea7661-b67f-4fbb-924a-943fe48995cb\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-6txpx" Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.410653 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2cf31abb-ef18-4ba5-8cc9-bf3f7507f771-config\") pod \"apiserver-76f77b778f-s6h9g\" (UID: \"2cf31abb-ef18-4ba5-8cc9-bf3f7507f771\") " pod="openshift-apiserver/apiserver-76f77b778f-s6h9g" Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.410671 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l4wbj\" (UniqueName: \"kubernetes.io/projected/a83bd6cc-62d8-44f4-930b-1c8fb70277c5-kube-api-access-l4wbj\") pod \"cluster-image-registry-operator-dc59b4c8b-9fg2s\" (UID: \"a83bd6cc-62d8-44f4-930b-1c8fb70277c5\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-9fg2s" Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.410689 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/2cf31abb-ef18-4ba5-8cc9-bf3f7507f771-etcd-serving-ca\") pod \"apiserver-76f77b778f-s6h9g\" (UID: \"2cf31abb-ef18-4ba5-8cc9-bf3f7507f771\") " pod="openshift-apiserver/apiserver-76f77b778f-s6h9g" Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.410705 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/c7db3f86-0e71-46d1-b95c-929069bf8b3a-client-ca\") pod \"route-controller-manager-6576b87f9c-2ph5v\" (UID: \"c7db3f86-0e71-46d1-b95c-929069bf8b3a\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-2ph5v" Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.410722 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/41587982-d7bb-413c-a1b4-12baa2ea600a-audit-policies\") pod \"apiserver-7bbb656c7d-66bs5\" (UID: \"41587982-d7bb-413c-a1b4-12baa2ea600a\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-66bs5" Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.410740 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/f3921e9a-9d9a-45e0-8733-6a1cd20422d2-etcd-service-ca\") pod \"etcd-operator-b45778765-p4gmt\" (UID: \"f3921e9a-9d9a-45e0-8733-6a1cd20422d2\") " pod="openshift-etcd-operator/etcd-operator-b45778765-p4gmt" Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.410759 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/2cf31abb-ef18-4ba5-8cc9-bf3f7507f771-trusted-ca-bundle\") pod \"apiserver-76f77b778f-s6h9g\" (UID: \"2cf31abb-ef18-4ba5-8cc9-bf3f7507f771\") " pod="openshift-apiserver/apiserver-76f77b778f-s6h9g" Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.410776 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q994j\" (UniqueName: \"kubernetes.io/projected/03939884-e3c9-47b9-a7cb-bc19e2f09c9c-kube-api-access-q994j\") pod \"cluster-samples-operator-665b6dd947-5zm4c\" (UID: \"03939884-e3c9-47b9-a7cb-bc19e2f09c9c\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-5zm4c" Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.410792 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f3921e9a-9d9a-45e0-8733-6a1cd20422d2-config\") pod \"etcd-operator-b45778765-p4gmt\" (UID: \"f3921e9a-9d9a-45e0-8733-6a1cd20422d2\") " pod="openshift-etcd-operator/etcd-operator-b45778765-p4gmt" Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.410808 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/41587982-d7bb-413c-a1b4-12baa2ea600a-etcd-client\") pod \"apiserver-7bbb656c7d-66bs5\" (UID: \"41587982-d7bb-413c-a1b4-12baa2ea600a\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-66bs5" Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.410878 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/2cf31abb-ef18-4ba5-8cc9-bf3f7507f771-audit-dir\") pod \"apiserver-76f77b778f-s6h9g\" (UID: \"2cf31abb-ef18-4ba5-8cc9-bf3f7507f771\") " pod="openshift-apiserver/apiserver-76f77b778f-s6h9g" Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.411255 4980 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-d6dd4"] Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.411667 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/2cf31abb-ef18-4ba5-8cc9-bf3f7507f771-node-pullsecrets\") pod \"apiserver-76f77b778f-s6h9g\" (UID: \"2cf31abb-ef18-4ba5-8cc9-bf3f7507f771\") " pod="openshift-apiserver/apiserver-76f77b778f-s6h9g" Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.411901 4980 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-8jc8f"] Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.412194 4980 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.412441 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-8jc8f" Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.412459 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/2cf31abb-ef18-4ba5-8cc9-bf3f7507f771-audit\") pod \"apiserver-76f77b778f-s6h9g\" (UID: \"2cf31abb-ef18-4ba5-8cc9-bf3f7507f771\") " pod="openshift-apiserver/apiserver-76f77b778f-s6h9g" Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.412704 4980 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-jp25g"] Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.412735 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-49749" Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.412771 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-d6dd4" Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.413389 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-jp25g" Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.413764 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2cf31abb-ef18-4ba5-8cc9-bf3f7507f771-config\") pod \"apiserver-76f77b778f-s6h9g\" (UID: \"2cf31abb-ef18-4ba5-8cc9-bf3f7507f771\") " pod="openshift-apiserver/apiserver-76f77b778f-s6h9g" Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.414435 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/2cf31abb-ef18-4ba5-8cc9-bf3f7507f771-etcd-serving-ca\") pod \"apiserver-76f77b778f-s6h9g\" (UID: \"2cf31abb-ef18-4ba5-8cc9-bf3f7507f771\") " pod="openshift-apiserver/apiserver-76f77b778f-s6h9g" Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.415479 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/2cf31abb-ef18-4ba5-8cc9-bf3f7507f771-trusted-ca-bundle\") pod \"apiserver-76f77b778f-s6h9g\" (UID: \"2cf31abb-ef18-4ba5-8cc9-bf3f7507f771\") " pod="openshift-apiserver/apiserver-76f77b778f-s6h9g" Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.416660 4980 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.423809 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/2cf31abb-ef18-4ba5-8cc9-bf3f7507f771-encryption-config\") pod \"apiserver-76f77b778f-s6h9g\" (UID: \"2cf31abb-ef18-4ba5-8cc9-bf3f7507f771\") " pod="openshift-apiserver/apiserver-76f77b778f-s6h9g" Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.423846 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2cf31abb-ef18-4ba5-8cc9-bf3f7507f771-serving-cert\") pod \"apiserver-76f77b778f-s6h9g\" (UID: \"2cf31abb-ef18-4ba5-8cc9-bf3f7507f771\") " pod="openshift-apiserver/apiserver-76f77b778f-s6h9g" Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.425802 4980 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.426074 4980 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.426255 4980 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.426442 4980 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.426552 4980 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.426596 4980 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.426666 4980 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.426855 4980 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.427036 4980 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.427245 4980 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.427415 4980 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.427649 4980 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.429743 4980 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-cqk6j"] Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.430468 4980 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.430702 4980 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.430797 4980 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.431462 4980 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.432872 4980 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-sq6bp"] Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.432980 4980 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.433171 4980 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.433271 4980 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.433355 4980 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.433443 4980 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.433480 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-sq6bp" Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.433555 4980 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.433638 4980 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.433726 4980 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.433816 4980 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.433892 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-cqk6j" Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.433918 4980 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.434021 4980 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.433355 4980 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-27gc7"] Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.434119 4980 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.434387 4980 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.434587 4980 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.434691 4980 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.434693 4980 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.434861 4980 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.458929 4980 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.459437 4980 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.459618 4980 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.459772 4980 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.461134 4980 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.462986 4980 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-tlpd8"] Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.463269 4980 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.464127 4980 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-cjvdn"] Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.465469 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-cjvdn" Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.466055 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-27gc7" Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.466211 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-tlpd8" Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.469741 4980 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-qcn9f"] Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.471065 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-qcn9f" Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.527617 4980 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-hjrkf"] Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.528729 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"f42b0cd3219b7939963face8d96ffe1557f3a42362a44deca4f0c3914ccbe8c0"} Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.528773 4980 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-56s8p"] Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.528870 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-hjrkf" Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.528934 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/443a2dc9-0546-4ab0-a448-eab81c09b482-oauth-serving-cert\") pod \"console-f9d7485db-gjmj8\" (UID: \"443a2dc9-0546-4ab0-a448-eab81c09b482\") " pod="openshift-console/console-f9d7485db-gjmj8" Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.528993 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n7rjm\" (UniqueName: \"kubernetes.io/projected/c7db3f86-0e71-46d1-b95c-929069bf8b3a-kube-api-access-n7rjm\") pod \"route-controller-manager-6576b87f9c-2ph5v\" (UID: \"c7db3f86-0e71-46d1-b95c-929069bf8b3a\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-2ph5v" Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.529029 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kdpcs\" (UniqueName: \"kubernetes.io/projected/005c8e70-2c2b-43a4-ae31-fd9013840824-kube-api-access-kdpcs\") pod \"machine-approver-56656f9798-425jl\" (UID: \"005c8e70-2c2b-43a4-ae31-fd9013840824\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-425jl" Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.529083 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ndb48\" (UniqueName: \"kubernetes.io/projected/f2ea7661-b67f-4fbb-924a-943fe48995cb-kube-api-access-ndb48\") pod \"openshift-apiserver-operator-796bbdcf4f-6txpx\" (UID: \"f2ea7661-b67f-4fbb-924a-943fe48995cb\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-6txpx" Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.529114 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/7c3cb795-33ac-4b16-b41e-a7da0de03578-available-featuregates\") pod \"openshift-config-operator-7777fb866f-zqsvr\" (UID: \"7c3cb795-33ac-4b16-b41e-a7da0de03578\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-zqsvr" Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.529169 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/f8973f62-160b-4c4b-9238-6a2ff3d63bc5-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-x5m4p\" (UID: \"f8973f62-160b-4c4b-9238-6a2ff3d63bc5\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-x5m4p" Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.529207 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/443a2dc9-0546-4ab0-a448-eab81c09b482-console-serving-cert\") pod \"console-f9d7485db-gjmj8\" (UID: \"443a2dc9-0546-4ab0-a448-eab81c09b482\") " pod="openshift-console/console-f9d7485db-gjmj8" Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.529257 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/443a2dc9-0546-4ab0-a448-eab81c09b482-console-oauth-config\") pod \"console-f9d7485db-gjmj8\" (UID: \"443a2dc9-0546-4ab0-a448-eab81c09b482\") " pod="openshift-console/console-f9d7485db-gjmj8" Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.529283 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f2ea7661-b67f-4fbb-924a-943fe48995cb-config\") pod \"openshift-apiserver-operator-796bbdcf4f-6txpx\" (UID: \"f2ea7661-b67f-4fbb-924a-943fe48995cb\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-6txpx" Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.529344 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l4wbj\" (UniqueName: \"kubernetes.io/projected/a83bd6cc-62d8-44f4-930b-1c8fb70277c5-kube-api-access-l4wbj\") pod \"cluster-image-registry-operator-dc59b4c8b-9fg2s\" (UID: \"a83bd6cc-62d8-44f4-930b-1c8fb70277c5\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-9fg2s" Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.529378 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/c7db3f86-0e71-46d1-b95c-929069bf8b3a-client-ca\") pod \"route-controller-manager-6576b87f9c-2ph5v\" (UID: \"c7db3f86-0e71-46d1-b95c-929069bf8b3a\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-2ph5v" Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.529431 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/41587982-d7bb-413c-a1b4-12baa2ea600a-audit-policies\") pod \"apiserver-7bbb656c7d-66bs5\" (UID: \"41587982-d7bb-413c-a1b4-12baa2ea600a\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-66bs5" Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.529455 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/f3921e9a-9d9a-45e0-8733-6a1cd20422d2-etcd-service-ca\") pod \"etcd-operator-b45778765-p4gmt\" (UID: \"f3921e9a-9d9a-45e0-8733-6a1cd20422d2\") " pod="openshift-etcd-operator/etcd-operator-b45778765-p4gmt" Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.529505 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q994j\" (UniqueName: \"kubernetes.io/projected/03939884-e3c9-47b9-a7cb-bc19e2f09c9c-kube-api-access-q994j\") pod \"cluster-samples-operator-665b6dd947-5zm4c\" (UID: \"03939884-e3c9-47b9-a7cb-bc19e2f09c9c\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-5zm4c" Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.529569 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f3921e9a-9d9a-45e0-8733-6a1cd20422d2-config\") pod \"etcd-operator-b45778765-p4gmt\" (UID: \"f3921e9a-9d9a-45e0-8733-6a1cd20422d2\") " pod="openshift-etcd-operator/etcd-operator-b45778765-p4gmt" Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.529594 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/41587982-d7bb-413c-a1b4-12baa2ea600a-etcd-client\") pod \"apiserver-7bbb656c7d-66bs5\" (UID: \"41587982-d7bb-413c-a1b4-12baa2ea600a\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-66bs5" Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.529685 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"452c441f3f24db96ee4c7585a33441e4206c9b9a2d573710d6f584e969c466dd"} Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.529719 4980 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.529774 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-56s8p" Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.530798 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/7c3cb795-33ac-4b16-b41e-a7da0de03578-available-featuregates\") pod \"openshift-config-operator-7777fb866f-zqsvr\" (UID: \"7c3cb795-33ac-4b16-b41e-a7da0de03578\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-zqsvr" Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.530946 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/443a2dc9-0546-4ab0-a448-eab81c09b482-oauth-serving-cert\") pod \"console-f9d7485db-gjmj8\" (UID: \"443a2dc9-0546-4ab0-a448-eab81c09b482\") " pod="openshift-console/console-f9d7485db-gjmj8" Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.530946 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/c7db3f86-0e71-46d1-b95c-929069bf8b3a-client-ca\") pod \"route-controller-manager-6576b87f9c-2ph5v\" (UID: \"c7db3f86-0e71-46d1-b95c-929069bf8b3a\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-2ph5v" Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.531568 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f2ea7661-b67f-4fbb-924a-943fe48995cb-config\") pod \"openshift-apiserver-operator-796bbdcf4f-6txpx\" (UID: \"f2ea7661-b67f-4fbb-924a-943fe48995cb\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-6txpx" Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.531696 4980 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.531732 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"154ab51b1e466e2a6b3eac6d12f56da87ee420481d93859db9921a38afee1248"} Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.531923 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"5d7eb89c42cd753080349aef30db6565512f9b5099c4595b53596f5cc3e52ae9"} Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.532098 4980 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.533964 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/2ac9a45e-ea4a-472e-bd5a-d947a0e2713e-client-ca\") pod \"controller-manager-879f6c89f-qpchc\" (UID: \"2ac9a45e-ea4a-472e-bd5a-d947a0e2713e\") " pod="openshift-controller-manager/controller-manager-879f6c89f-qpchc" Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.536408 4980 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.536687 4980 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-qpchc"] Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.537483 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"069c041dc72b7671fd354ca2f65268ccbace0af40b68e3f07e75b812cec61beb"} Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.537542 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"3a9aa0575b5feb4c551547923bf874e17853a47d7e530921747c6791df8c23fe"} Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.538616 4980 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.538881 4980 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-2ph5v"] Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.539174 4980 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.539396 4980 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-s6h9g"] Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.539680 4980 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.539800 4980 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.539804 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/41587982-d7bb-413c-a1b4-12baa2ea600a-audit-policies\") pod \"apiserver-7bbb656c7d-66bs5\" (UID: \"41587982-d7bb-413c-a1b4-12baa2ea600a\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-66bs5" Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.529645 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/2ac9a45e-ea4a-472e-bd5a-d947a0e2713e-client-ca\") pod \"controller-manager-879f6c89f-qpchc\" (UID: \"2ac9a45e-ea4a-472e-bd5a-d947a0e2713e\") " pod="openshift-controller-manager/controller-manager-879f6c89f-qpchc" Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.539984 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/005c8e70-2c2b-43a4-ae31-fd9013840824-auth-proxy-config\") pod \"machine-approver-56656f9798-425jl\" (UID: \"005c8e70-2c2b-43a4-ae31-fd9013840824\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-425jl" Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.540144 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f3921e9a-9d9a-45e0-8733-6a1cd20422d2-serving-cert\") pod \"etcd-operator-b45778765-p4gmt\" (UID: \"f3921e9a-9d9a-45e0-8733-6a1cd20422d2\") " pod="openshift-etcd-operator/etcd-operator-b45778765-p4gmt" Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.540189 4980 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.540214 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/443a2dc9-0546-4ab0-a448-eab81c09b482-console-config\") pod \"console-f9d7485db-gjmj8\" (UID: \"443a2dc9-0546-4ab0-a448-eab81c09b482\") " pod="openshift-console/console-f9d7485db-gjmj8" Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.540245 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f2ea7661-b67f-4fbb-924a-943fe48995cb-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-6txpx\" (UID: \"f2ea7661-b67f-4fbb-924a-943fe48995cb\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-6txpx" Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.540300 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f8973f62-160b-4c4b-9238-6a2ff3d63bc5-config\") pod \"machine-api-operator-5694c8668f-x5m4p\" (UID: \"f8973f62-160b-4c4b-9238-6a2ff3d63bc5\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-x5m4p" Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.540332 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a83bd6cc-62d8-44f4-930b-1c8fb70277c5-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-9fg2s\" (UID: \"a83bd6cc-62d8-44f4-930b-1c8fb70277c5\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-9fg2s" Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.540398 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2ac9a45e-ea4a-472e-bd5a-d947a0e2713e-config\") pod \"controller-manager-879f6c89f-qpchc\" (UID: \"2ac9a45e-ea4a-472e-bd5a-d947a0e2713e\") " pod="openshift-controller-manager/controller-manager-879f6c89f-qpchc" Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.540464 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/41587982-d7bb-413c-a1b4-12baa2ea600a-encryption-config\") pod \"apiserver-7bbb656c7d-66bs5\" (UID: \"41587982-d7bb-413c-a1b4-12baa2ea600a\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-66bs5" Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.540531 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/41587982-d7bb-413c-a1b4-12baa2ea600a-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-66bs5\" (UID: \"41587982-d7bb-413c-a1b4-12baa2ea600a\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-66bs5" Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.540569 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dqls7\" (UniqueName: \"kubernetes.io/projected/41587982-d7bb-413c-a1b4-12baa2ea600a-kube-api-access-dqls7\") pod \"apiserver-7bbb656c7d-66bs5\" (UID: \"41587982-d7bb-413c-a1b4-12baa2ea600a\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-66bs5" Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.540620 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/443a2dc9-0546-4ab0-a448-eab81c09b482-service-ca\") pod \"console-f9d7485db-gjmj8\" (UID: \"443a2dc9-0546-4ab0-a448-eab81c09b482\") " pod="openshift-console/console-f9d7485db-gjmj8" Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.540649 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c7db3f86-0e71-46d1-b95c-929069bf8b3a-serving-cert\") pod \"route-controller-manager-6576b87f9c-2ph5v\" (UID: \"c7db3f86-0e71-46d1-b95c-929069bf8b3a\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-2ph5v" Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.540706 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/443a2dc9-0546-4ab0-a448-eab81c09b482-trusted-ca-bundle\") pod \"console-f9d7485db-gjmj8\" (UID: \"443a2dc9-0546-4ab0-a448-eab81c09b482\") " pod="openshift-console/console-f9d7485db-gjmj8" Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.540739 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-svpft\" (UniqueName: \"kubernetes.io/projected/443a2dc9-0546-4ab0-a448-eab81c09b482-kube-api-access-svpft\") pod \"console-f9d7485db-gjmj8\" (UID: \"443a2dc9-0546-4ab0-a448-eab81c09b482\") " pod="openshift-console/console-f9d7485db-gjmj8" Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.540789 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/f3921e9a-9d9a-45e0-8733-6a1cd20422d2-etcd-client\") pod \"etcd-operator-b45778765-p4gmt\" (UID: \"f3921e9a-9d9a-45e0-8733-6a1cd20422d2\") " pod="openshift-etcd-operator/etcd-operator-b45778765-p4gmt" Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.540823 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/41587982-d7bb-413c-a1b4-12baa2ea600a-serving-cert\") pod \"apiserver-7bbb656c7d-66bs5\" (UID: \"41587982-d7bb-413c-a1b4-12baa2ea600a\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-66bs5" Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.540876 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/41587982-d7bb-413c-a1b4-12baa2ea600a-audit-dir\") pod \"apiserver-7bbb656c7d-66bs5\" (UID: \"41587982-d7bb-413c-a1b4-12baa2ea600a\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-66bs5" Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.540902 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7c3cb795-33ac-4b16-b41e-a7da0de03578-serving-cert\") pod \"openshift-config-operator-7777fb866f-zqsvr\" (UID: \"7c3cb795-33ac-4b16-b41e-a7da0de03578\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-zqsvr" Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.540955 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dfhqz\" (UniqueName: \"kubernetes.io/projected/7c3cb795-33ac-4b16-b41e-a7da0de03578-kube-api-access-dfhqz\") pod \"openshift-config-operator-7777fb866f-zqsvr\" (UID: \"7c3cb795-33ac-4b16-b41e-a7da0de03578\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-zqsvr" Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.540987 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/f8973f62-160b-4c4b-9238-6a2ff3d63bc5-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-x5m4p\" (UID: \"f8973f62-160b-4c4b-9238-6a2ff3d63bc5\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-x5m4p" Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.540987 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/a83bd6cc-62d8-44f4-930b-1c8fb70277c5-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-9fg2s\" (UID: \"a83bd6cc-62d8-44f4-930b-1c8fb70277c5\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-9fg2s" Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.541060 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a83bd6cc-62d8-44f4-930b-1c8fb70277c5-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-9fg2s\" (UID: \"a83bd6cc-62d8-44f4-930b-1c8fb70277c5\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-9fg2s" Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.541094 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/005c8e70-2c2b-43a4-ae31-fd9013840824-config\") pod \"machine-approver-56656f9798-425jl\" (UID: \"005c8e70-2c2b-43a4-ae31-fd9013840824\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-425jl" Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.541119 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dbbbf\" (UniqueName: \"kubernetes.io/projected/f8973f62-160b-4c4b-9238-6a2ff3d63bc5-kube-api-access-dbbbf\") pod \"machine-api-operator-5694c8668f-x5m4p\" (UID: \"f8973f62-160b-4c4b-9238-6a2ff3d63bc5\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-x5m4p" Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.541141 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/41587982-d7bb-413c-a1b4-12baa2ea600a-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-66bs5\" (UID: \"41587982-d7bb-413c-a1b4-12baa2ea600a\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-66bs5" Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.541151 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wmtnf\" (UniqueName: \"kubernetes.io/projected/f3921e9a-9d9a-45e0-8733-6a1cd20422d2-kube-api-access-wmtnf\") pod \"etcd-operator-b45778765-p4gmt\" (UID: \"f3921e9a-9d9a-45e0-8733-6a1cd20422d2\") " pod="openshift-etcd-operator/etcd-operator-b45778765-p4gmt" Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.541180 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/41587982-d7bb-413c-a1b4-12baa2ea600a-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-66bs5\" (UID: \"41587982-d7bb-413c-a1b4-12baa2ea600a\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-66bs5" Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.541206 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/03939884-e3c9-47b9-a7cb-bc19e2f09c9c-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-5zm4c\" (UID: \"03939884-e3c9-47b9-a7cb-bc19e2f09c9c\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-5zm4c" Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.541243 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/2ac9a45e-ea4a-472e-bd5a-d947a0e2713e-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-qpchc\" (UID: \"2ac9a45e-ea4a-472e-bd5a-d947a0e2713e\") " pod="openshift-controller-manager/controller-manager-879f6c89f-qpchc" Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.541268 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c7db3f86-0e71-46d1-b95c-929069bf8b3a-config\") pod \"route-controller-manager-6576b87f9c-2ph5v\" (UID: \"c7db3f86-0e71-46d1-b95c-929069bf8b3a\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-2ph5v" Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.541309 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/005c8e70-2c2b-43a4-ae31-fd9013840824-machine-approver-tls\") pod \"machine-approver-56656f9798-425jl\" (UID: \"005c8e70-2c2b-43a4-ae31-fd9013840824\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-425jl" Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.541336 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2ac9a45e-ea4a-472e-bd5a-d947a0e2713e-serving-cert\") pod \"controller-manager-879f6c89f-qpchc\" (UID: \"2ac9a45e-ea4a-472e-bd5a-d947a0e2713e\") " pod="openshift-controller-manager/controller-manager-879f6c89f-qpchc" Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.541383 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/f8973f62-160b-4c4b-9238-6a2ff3d63bc5-images\") pod \"machine-api-operator-5694c8668f-x5m4p\" (UID: \"f8973f62-160b-4c4b-9238-6a2ff3d63bc5\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-x5m4p" Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.541439 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r85l6\" (UniqueName: \"kubernetes.io/projected/2ac9a45e-ea4a-472e-bd5a-d947a0e2713e-kube-api-access-r85l6\") pod \"controller-manager-879f6c89f-qpchc\" (UID: \"2ac9a45e-ea4a-472e-bd5a-d947a0e2713e\") " pod="openshift-controller-manager/controller-manager-879f6c89f-qpchc" Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.541451 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/443a2dc9-0546-4ab0-a448-eab81c09b482-console-config\") pod \"console-f9d7485db-gjmj8\" (UID: \"443a2dc9-0546-4ab0-a448-eab81c09b482\") " pod="openshift-console/console-f9d7485db-gjmj8" Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.541464 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/f3921e9a-9d9a-45e0-8733-6a1cd20422d2-etcd-ca\") pod \"etcd-operator-b45778765-p4gmt\" (UID: \"f3921e9a-9d9a-45e0-8733-6a1cd20422d2\") " pod="openshift-etcd-operator/etcd-operator-b45778765-p4gmt" Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.542622 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/443a2dc9-0546-4ab0-a448-eab81c09b482-service-ca\") pod \"console-f9d7485db-gjmj8\" (UID: \"443a2dc9-0546-4ab0-a448-eab81c09b482\") " pod="openshift-console/console-f9d7485db-gjmj8" Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.542754 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/443a2dc9-0546-4ab0-a448-eab81c09b482-console-serving-cert\") pod \"console-f9d7485db-gjmj8\" (UID: \"443a2dc9-0546-4ab0-a448-eab81c09b482\") " pod="openshift-console/console-f9d7485db-gjmj8" Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.542846 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/41587982-d7bb-413c-a1b4-12baa2ea600a-audit-dir\") pod \"apiserver-7bbb656c7d-66bs5\" (UID: \"41587982-d7bb-413c-a1b4-12baa2ea600a\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-66bs5" Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.546605 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a83bd6cc-62d8-44f4-930b-1c8fb70277c5-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-9fg2s\" (UID: \"a83bd6cc-62d8-44f4-930b-1c8fb70277c5\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-9fg2s" Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.547138 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/2ac9a45e-ea4a-472e-bd5a-d947a0e2713e-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-qpchc\" (UID: \"2ac9a45e-ea4a-472e-bd5a-d947a0e2713e\") " pod="openshift-controller-manager/controller-manager-879f6c89f-qpchc" Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.547853 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/005c8e70-2c2b-43a4-ae31-fd9013840824-config\") pod \"machine-approver-56656f9798-425jl\" (UID: \"005c8e70-2c2b-43a4-ae31-fd9013840824\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-425jl" Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.548994 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/41587982-d7bb-413c-a1b4-12baa2ea600a-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-66bs5\" (UID: \"41587982-d7bb-413c-a1b4-12baa2ea600a\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-66bs5" Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.550742 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/443a2dc9-0546-4ab0-a448-eab81c09b482-console-oauth-config\") pod \"console-f9d7485db-gjmj8\" (UID: \"443a2dc9-0546-4ab0-a448-eab81c09b482\") " pod="openshift-console/console-f9d7485db-gjmj8" Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.552079 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f8973f62-160b-4c4b-9238-6a2ff3d63bc5-config\") pod \"machine-api-operator-5694c8668f-x5m4p\" (UID: \"f8973f62-160b-4c4b-9238-6a2ff3d63bc5\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-x5m4p" Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.552180 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7c3cb795-33ac-4b16-b41e-a7da0de03578-serving-cert\") pod \"openshift-config-operator-7777fb866f-zqsvr\" (UID: \"7c3cb795-33ac-4b16-b41e-a7da0de03578\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-zqsvr" Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.553350 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2ac9a45e-ea4a-472e-bd5a-d947a0e2713e-config\") pod \"controller-manager-879f6c89f-qpchc\" (UID: \"2ac9a45e-ea4a-472e-bd5a-d947a0e2713e\") " pod="openshift-controller-manager/controller-manager-879f6c89f-qpchc" Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.540710 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/005c8e70-2c2b-43a4-ae31-fd9013840824-auth-proxy-config\") pod \"machine-approver-56656f9798-425jl\" (UID: \"005c8e70-2c2b-43a4-ae31-fd9013840824\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-425jl" Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.561058 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c7db3f86-0e71-46d1-b95c-929069bf8b3a-config\") pod \"route-controller-manager-6576b87f9c-2ph5v\" (UID: \"c7db3f86-0e71-46d1-b95c-929069bf8b3a\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-2ph5v" Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.561331 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/f8973f62-160b-4c4b-9238-6a2ff3d63bc5-images\") pod \"machine-api-operator-5694c8668f-x5m4p\" (UID: \"f8973f62-160b-4c4b-9238-6a2ff3d63bc5\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-x5m4p" Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.561501 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/443a2dc9-0546-4ab0-a448-eab81c09b482-trusted-ca-bundle\") pod \"console-f9d7485db-gjmj8\" (UID: \"443a2dc9-0546-4ab0-a448-eab81c09b482\") " pod="openshift-console/console-f9d7485db-gjmj8" Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.561696 4980 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-spd2v"] Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.562298 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/a83bd6cc-62d8-44f4-930b-1c8fb70277c5-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-9fg2s\" (UID: \"a83bd6cc-62d8-44f4-930b-1c8fb70277c5\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-9fg2s" Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.562687 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-spd2v" Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.563734 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/005c8e70-2c2b-43a4-ae31-fd9013840824-machine-approver-tls\") pod \"machine-approver-56656f9798-425jl\" (UID: \"005c8e70-2c2b-43a4-ae31-fd9013840824\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-425jl" Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.563819 4980 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-r47r5"] Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.564157 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2ac9a45e-ea4a-472e-bd5a-d947a0e2713e-serving-cert\") pod \"controller-manager-879f6c89f-qpchc\" (UID: \"2ac9a45e-ea4a-472e-bd5a-d947a0e2713e\") " pod="openshift-controller-manager/controller-manager-879f6c89f-qpchc" Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.564956 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-r47r5" Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.567248 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c7db3f86-0e71-46d1-b95c-929069bf8b3a-serving-cert\") pod \"route-controller-manager-6576b87f9c-2ph5v\" (UID: \"c7db3f86-0e71-46d1-b95c-929069bf8b3a\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-2ph5v" Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.566556 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/41587982-d7bb-413c-a1b4-12baa2ea600a-serving-cert\") pod \"apiserver-7bbb656c7d-66bs5\" (UID: \"41587982-d7bb-413c-a1b4-12baa2ea600a\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-66bs5" Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.565891 4980 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29416530-9hm9m"] Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.567591 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/41587982-d7bb-413c-a1b4-12baa2ea600a-etcd-client\") pod \"apiserver-7bbb656c7d-66bs5\" (UID: \"41587982-d7bb-413c-a1b4-12baa2ea600a\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-66bs5" Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.567766 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/41587982-d7bb-413c-a1b4-12baa2ea600a-encryption-config\") pod \"apiserver-7bbb656c7d-66bs5\" (UID: \"41587982-d7bb-413c-a1b4-12baa2ea600a\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-66bs5" Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.568321 4980 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-qpf6b"] Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.568424 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29416530-9hm9m" Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.568976 4980 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-gjdsq"] Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.569065 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-qpf6b" Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.569701 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-gjdsq" Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.570370 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/03939884-e3c9-47b9-a7cb-bc19e2f09c9c-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-5zm4c\" (UID: \"03939884-e3c9-47b9-a7cb-bc19e2f09c9c\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-5zm4c" Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.573842 4980 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.574292 4980 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/dns-default-4dcjh"] Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.575229 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-4dcjh" Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.577226 4980 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-xjtnv"] Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.578766 4980 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-9fg2s"] Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.579825 4980 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.580032 4980 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-zqsvr"] Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.580370 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f2ea7661-b67f-4fbb-924a-943fe48995cb-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-6txpx\" (UID: \"f2ea7661-b67f-4fbb-924a-943fe48995cb\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-6txpx" Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.582036 4980 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-x5m4p"] Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.585481 4980 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-6txpx"] Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.592022 4980 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-w6bmg"] Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.593092 4980 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-84ft2"] Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.594167 4980 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-6h96g"] Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.595344 4980 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-66bs5"] Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.595689 4980 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.596718 4980 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-gjmj8"] Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.598156 4980 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-qw2fb"] Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.599688 4980 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-tlpd8"] Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.601823 4980 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-49749"] Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.601893 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f3921e9a-9d9a-45e0-8733-6a1cd20422d2-config\") pod \"etcd-operator-b45778765-p4gmt\" (UID: \"f3921e9a-9d9a-45e0-8733-6a1cd20422d2\") " pod="openshift-etcd-operator/etcd-operator-b45778765-p4gmt" Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.603654 4980 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-cqk6j"] Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.604968 4980 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-cjvdn"] Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.606052 4980 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-spd2v"] Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.607512 4980 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-canary/ingress-canary-799j5"] Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.608897 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-799j5" Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.609001 4980 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-74hx2"] Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.612359 4980 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.612649 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-74hx2" Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.615026 4980 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-27gc7"] Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.620546 4980 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-jp25g"] Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.624346 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f3921e9a-9d9a-45e0-8733-6a1cd20422d2-serving-cert\") pod \"etcd-operator-b45778765-p4gmt\" (UID: \"f3921e9a-9d9a-45e0-8733-6a1cd20422d2\") " pod="openshift-etcd-operator/etcd-operator-b45778765-p4gmt" Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.624957 4980 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-pmmcw"] Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.626215 4980 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-sq6bp"] Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.627363 4980 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-5zm4c"] Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.630446 4980 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-lrnhm"] Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.630752 4980 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.630954 4980 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-d6dd4"] Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.632774 4980 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-p4gmt"] Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.634142 4980 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-qcn9f"] Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.635550 4980 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-kftgf"] Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.636266 4980 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-hjrkf"] Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.637346 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/f3921e9a-9d9a-45e0-8733-6a1cd20422d2-etcd-client\") pod \"etcd-operator-b45778765-p4gmt\" (UID: \"f3921e9a-9d9a-45e0-8733-6a1cd20422d2\") " pod="openshift-etcd-operator/etcd-operator-b45778765-p4gmt" Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.637401 4980 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-8jc8f"] Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.639714 4980 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-4dcjh"] Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.641848 4980 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-56s8p"] Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.643640 4980 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29416530-9hm9m"] Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.645556 4980 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-qpf6b"] Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.646091 4980 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-799j5"] Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.647203 4980 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-r47r5"] Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.648740 4980 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-74hx2"] Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.650353 4980 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.650976 4980 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-gjdsq"] Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.654253 4980 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-server-bfp7v"] Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.654999 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-bfp7v" Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.670064 4980 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.672354 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/f3921e9a-9d9a-45e0-8733-6a1cd20422d2-etcd-ca\") pod \"etcd-operator-b45778765-p4gmt\" (UID: \"f3921e9a-9d9a-45e0-8733-6a1cd20422d2\") " pod="openshift-etcd-operator/etcd-operator-b45778765-p4gmt" Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.695244 4980 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.700879 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/f3921e9a-9d9a-45e0-8733-6a1cd20422d2-etcd-service-ca\") pod \"etcd-operator-b45778765-p4gmt\" (UID: \"f3921e9a-9d9a-45e0-8733-6a1cd20422d2\") " pod="openshift-etcd-operator/etcd-operator-b45778765-p4gmt" Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.710144 4980 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.729272 4980 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.769870 4980 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.790734 4980 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.809928 4980 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.830428 4980 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.849906 4980 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.870619 4980 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.890015 4980 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.910576 4980 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.931083 4980 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.950155 4980 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.970118 4980 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Dec 06 03:36:31 crc kubenswrapper[4980]: I1206 03:36:31.990548 4980 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Dec 06 03:36:32 crc kubenswrapper[4980]: I1206 03:36:32.009635 4980 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Dec 06 03:36:32 crc kubenswrapper[4980]: I1206 03:36:32.030163 4980 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Dec 06 03:36:32 crc kubenswrapper[4980]: I1206 03:36:32.053675 4980 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Dec 06 03:36:32 crc kubenswrapper[4980]: I1206 03:36:32.070134 4980 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Dec 06 03:36:32 crc kubenswrapper[4980]: I1206 03:36:32.090770 4980 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Dec 06 03:36:32 crc kubenswrapper[4980]: I1206 03:36:32.103315 4980 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-5w4fr" Dec 06 03:36:32 crc kubenswrapper[4980]: I1206 03:36:32.109555 4980 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Dec 06 03:36:32 crc kubenswrapper[4980]: I1206 03:36:32.130535 4980 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Dec 06 03:36:32 crc kubenswrapper[4980]: I1206 03:36:32.151049 4980 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Dec 06 03:36:32 crc kubenswrapper[4980]: I1206 03:36:32.170635 4980 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Dec 06 03:36:32 crc kubenswrapper[4980]: I1206 03:36:32.189981 4980 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Dec 06 03:36:32 crc kubenswrapper[4980]: I1206 03:36:32.209667 4980 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Dec 06 03:36:32 crc kubenswrapper[4980]: I1206 03:36:32.230853 4980 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Dec 06 03:36:32 crc kubenswrapper[4980]: I1206 03:36:32.270852 4980 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Dec 06 03:36:32 crc kubenswrapper[4980]: I1206 03:36:32.286290 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6l7xr\" (UniqueName: \"kubernetes.io/projected/2cf31abb-ef18-4ba5-8cc9-bf3f7507f771-kube-api-access-6l7xr\") pod \"apiserver-76f77b778f-s6h9g\" (UID: \"2cf31abb-ef18-4ba5-8cc9-bf3f7507f771\") " pod="openshift-apiserver/apiserver-76f77b778f-s6h9g" Dec 06 03:36:32 crc kubenswrapper[4980]: I1206 03:36:32.290413 4980 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Dec 06 03:36:32 crc kubenswrapper[4980]: I1206 03:36:32.310772 4980 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Dec 06 03:36:32 crc kubenswrapper[4980]: I1206 03:36:32.330851 4980 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Dec 06 03:36:32 crc kubenswrapper[4980]: I1206 03:36:32.352318 4980 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Dec 06 03:36:32 crc kubenswrapper[4980]: I1206 03:36:32.372491 4980 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Dec 06 03:36:32 crc kubenswrapper[4980]: I1206 03:36:32.393159 4980 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Dec 06 03:36:32 crc kubenswrapper[4980]: E1206 03:36:32.412582 4980 configmap.go:193] Couldn't get configMap openshift-apiserver/image-import-ca: failed to sync configmap cache: timed out waiting for the condition Dec 06 03:36:32 crc kubenswrapper[4980]: E1206 03:36:32.412670 4980 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/2cf31abb-ef18-4ba5-8cc9-bf3f7507f771-image-import-ca podName:2cf31abb-ef18-4ba5-8cc9-bf3f7507f771 nodeName:}" failed. No retries permitted until 2025-12-06 03:36:32.912651979 +0000 UTC m=+152.157403240 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "image-import-ca" (UniqueName: "kubernetes.io/configmap/2cf31abb-ef18-4ba5-8cc9-bf3f7507f771-image-import-ca") pod "apiserver-76f77b778f-s6h9g" (UID: "2cf31abb-ef18-4ba5-8cc9-bf3f7507f771") : failed to sync configmap cache: timed out waiting for the condition Dec 06 03:36:32 crc kubenswrapper[4980]: E1206 03:36:32.412582 4980 secret.go:188] Couldn't get secret openshift-apiserver/etcd-client: failed to sync secret cache: timed out waiting for the condition Dec 06 03:36:32 crc kubenswrapper[4980]: E1206 03:36:32.412862 4980 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/2cf31abb-ef18-4ba5-8cc9-bf3f7507f771-etcd-client podName:2cf31abb-ef18-4ba5-8cc9-bf3f7507f771 nodeName:}" failed. No retries permitted until 2025-12-06 03:36:32.912854274 +0000 UTC m=+152.157605545 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "etcd-client" (UniqueName: "kubernetes.io/secret/2cf31abb-ef18-4ba5-8cc9-bf3f7507f771-etcd-client") pod "apiserver-76f77b778f-s6h9g" (UID: "2cf31abb-ef18-4ba5-8cc9-bf3f7507f771") : failed to sync secret cache: timed out waiting for the condition Dec 06 03:36:32 crc kubenswrapper[4980]: I1206 03:36:32.429118 4980 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Dec 06 03:36:32 crc kubenswrapper[4980]: I1206 03:36:32.449357 4980 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Dec 06 03:36:32 crc kubenswrapper[4980]: I1206 03:36:32.468055 4980 request.go:700] Waited for 1.001618585s due to client-side throttling, not priority and fairness, request: GET:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-multus/secrets?fieldSelector=metadata.name%3Dmultus-admission-controller-secret&limit=500&resourceVersion=0 Dec 06 03:36:32 crc kubenswrapper[4980]: I1206 03:36:32.469224 4980 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Dec 06 03:36:32 crc kubenswrapper[4980]: I1206 03:36:32.490609 4980 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Dec 06 03:36:32 crc kubenswrapper[4980]: I1206 03:36:32.512709 4980 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Dec 06 03:36:32 crc kubenswrapper[4980]: I1206 03:36:32.530045 4980 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Dec 06 03:36:32 crc kubenswrapper[4980]: I1206 03:36:32.550741 4980 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Dec 06 03:36:32 crc kubenswrapper[4980]: I1206 03:36:32.570866 4980 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Dec 06 03:36:32 crc kubenswrapper[4980]: I1206 03:36:32.591963 4980 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Dec 06 03:36:32 crc kubenswrapper[4980]: I1206 03:36:32.622487 4980 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Dec 06 03:36:32 crc kubenswrapper[4980]: I1206 03:36:32.630798 4980 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Dec 06 03:36:32 crc kubenswrapper[4980]: I1206 03:36:32.649648 4980 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Dec 06 03:36:32 crc kubenswrapper[4980]: I1206 03:36:32.671257 4980 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Dec 06 03:36:32 crc kubenswrapper[4980]: I1206 03:36:32.690445 4980 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Dec 06 03:36:32 crc kubenswrapper[4980]: I1206 03:36:32.730725 4980 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Dec 06 03:36:32 crc kubenswrapper[4980]: I1206 03:36:32.739727 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l4wbj\" (UniqueName: \"kubernetes.io/projected/a83bd6cc-62d8-44f4-930b-1c8fb70277c5-kube-api-access-l4wbj\") pod \"cluster-image-registry-operator-dc59b4c8b-9fg2s\" (UID: \"a83bd6cc-62d8-44f4-930b-1c8fb70277c5\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-9fg2s" Dec 06 03:36:32 crc kubenswrapper[4980]: I1206 03:36:32.750114 4980 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Dec 06 03:36:32 crc kubenswrapper[4980]: I1206 03:36:32.770469 4980 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Dec 06 03:36:32 crc kubenswrapper[4980]: I1206 03:36:32.814496 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kdpcs\" (UniqueName: \"kubernetes.io/projected/005c8e70-2c2b-43a4-ae31-fd9013840824-kube-api-access-kdpcs\") pod \"machine-approver-56656f9798-425jl\" (UID: \"005c8e70-2c2b-43a4-ae31-fd9013840824\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-425jl" Dec 06 03:36:32 crc kubenswrapper[4980]: I1206 03:36:32.840556 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-425jl" Dec 06 03:36:32 crc kubenswrapper[4980]: I1206 03:36:32.840535 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ndb48\" (UniqueName: \"kubernetes.io/projected/f2ea7661-b67f-4fbb-924a-943fe48995cb-kube-api-access-ndb48\") pod \"openshift-apiserver-operator-796bbdcf4f-6txpx\" (UID: \"f2ea7661-b67f-4fbb-924a-943fe48995cb\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-6txpx" Dec 06 03:36:32 crc kubenswrapper[4980]: I1206 03:36:32.844823 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-6txpx" Dec 06 03:36:32 crc kubenswrapper[4980]: I1206 03:36:32.845803 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q994j\" (UniqueName: \"kubernetes.io/projected/03939884-e3c9-47b9-a7cb-bc19e2f09c9c-kube-api-access-q994j\") pod \"cluster-samples-operator-665b6dd947-5zm4c\" (UID: \"03939884-e3c9-47b9-a7cb-bc19e2f09c9c\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-5zm4c" Dec 06 03:36:32 crc kubenswrapper[4980]: I1206 03:36:32.865966 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n7rjm\" (UniqueName: \"kubernetes.io/projected/c7db3f86-0e71-46d1-b95c-929069bf8b3a-kube-api-access-n7rjm\") pod \"route-controller-manager-6576b87f9c-2ph5v\" (UID: \"c7db3f86-0e71-46d1-b95c-929069bf8b3a\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-2ph5v" Dec 06 03:36:32 crc kubenswrapper[4980]: I1206 03:36:32.891268 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-svpft\" (UniqueName: \"kubernetes.io/projected/443a2dc9-0546-4ab0-a448-eab81c09b482-kube-api-access-svpft\") pod \"console-f9d7485db-gjmj8\" (UID: \"443a2dc9-0546-4ab0-a448-eab81c09b482\") " pod="openshift-console/console-f9d7485db-gjmj8" Dec 06 03:36:32 crc kubenswrapper[4980]: I1206 03:36:32.903169 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-gjmj8" Dec 06 03:36:32 crc kubenswrapper[4980]: I1206 03:36:32.905235 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dqls7\" (UniqueName: \"kubernetes.io/projected/41587982-d7bb-413c-a1b4-12baa2ea600a-kube-api-access-dqls7\") pod \"apiserver-7bbb656c7d-66bs5\" (UID: \"41587982-d7bb-413c-a1b4-12baa2ea600a\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-66bs5" Dec 06 03:36:32 crc kubenswrapper[4980]: I1206 03:36:32.924595 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dbbbf\" (UniqueName: \"kubernetes.io/projected/f8973f62-160b-4c4b-9238-6a2ff3d63bc5-kube-api-access-dbbbf\") pod \"machine-api-operator-5694c8668f-x5m4p\" (UID: \"f8973f62-160b-4c4b-9238-6a2ff3d63bc5\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-x5m4p" Dec 06 03:36:32 crc kubenswrapper[4980]: I1206 03:36:32.930721 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-5zm4c" Dec 06 03:36:32 crc kubenswrapper[4980]: I1206 03:36:32.967337 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-2ph5v" Dec 06 03:36:32 crc kubenswrapper[4980]: I1206 03:36:32.969379 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/2cf31abb-ef18-4ba5-8cc9-bf3f7507f771-image-import-ca\") pod \"apiserver-76f77b778f-s6h9g\" (UID: \"2cf31abb-ef18-4ba5-8cc9-bf3f7507f771\") " pod="openshift-apiserver/apiserver-76f77b778f-s6h9g" Dec 06 03:36:32 crc kubenswrapper[4980]: I1206 03:36:32.969433 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/2cf31abb-ef18-4ba5-8cc9-bf3f7507f771-etcd-client\") pod \"apiserver-76f77b778f-s6h9g\" (UID: \"2cf31abb-ef18-4ba5-8cc9-bf3f7507f771\") " pod="openshift-apiserver/apiserver-76f77b778f-s6h9g" Dec 06 03:36:33 crc kubenswrapper[4980]: I1206 03:36:33.017673 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dfhqz\" (UniqueName: \"kubernetes.io/projected/7c3cb795-33ac-4b16-b41e-a7da0de03578-kube-api-access-dfhqz\") pod \"openshift-config-operator-7777fb866f-zqsvr\" (UID: \"7c3cb795-33ac-4b16-b41e-a7da0de03578\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-zqsvr" Dec 06 03:36:33 crc kubenswrapper[4980]: I1206 03:36:33.018617 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wmtnf\" (UniqueName: \"kubernetes.io/projected/f3921e9a-9d9a-45e0-8733-6a1cd20422d2-kube-api-access-wmtnf\") pod \"etcd-operator-b45778765-p4gmt\" (UID: \"f3921e9a-9d9a-45e0-8733-6a1cd20422d2\") " pod="openshift-etcd-operator/etcd-operator-b45778765-p4gmt" Dec 06 03:36:33 crc kubenswrapper[4980]: I1206 03:36:33.018646 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a83bd6cc-62d8-44f4-930b-1c8fb70277c5-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-9fg2s\" (UID: \"a83bd6cc-62d8-44f4-930b-1c8fb70277c5\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-9fg2s" Dec 06 03:36:33 crc kubenswrapper[4980]: I1206 03:36:33.021121 4980 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Dec 06 03:36:33 crc kubenswrapper[4980]: I1206 03:36:33.028571 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r85l6\" (UniqueName: \"kubernetes.io/projected/2ac9a45e-ea4a-472e-bd5a-d947a0e2713e-kube-api-access-r85l6\") pod \"controller-manager-879f6c89f-qpchc\" (UID: \"2ac9a45e-ea4a-472e-bd5a-d947a0e2713e\") " pod="openshift-controller-manager/controller-manager-879f6c89f-qpchc" Dec 06 03:36:33 crc kubenswrapper[4980]: I1206 03:36:33.028741 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-x5m4p" Dec 06 03:36:33 crc kubenswrapper[4980]: I1206 03:36:33.036628 4980 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Dec 06 03:36:33 crc kubenswrapper[4980]: I1206 03:36:33.110386 4980 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Dec 06 03:36:33 crc kubenswrapper[4980]: I1206 03:36:33.129992 4980 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Dec 06 03:36:33 crc kubenswrapper[4980]: I1206 03:36:33.130755 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-zqsvr" Dec 06 03:36:33 crc kubenswrapper[4980]: I1206 03:36:33.130770 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-9fg2s" Dec 06 03:36:33 crc kubenswrapper[4980]: I1206 03:36:33.132936 4980 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Dec 06 03:36:33 crc kubenswrapper[4980]: I1206 03:36:33.132950 4980 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Dec 06 03:36:33 crc kubenswrapper[4980]: I1206 03:36:33.132994 4980 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Dec 06 03:36:33 crc kubenswrapper[4980]: I1206 03:36:33.133142 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-qpchc" Dec 06 03:36:33 crc kubenswrapper[4980]: I1206 03:36:33.134454 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-66bs5" Dec 06 03:36:33 crc kubenswrapper[4980]: I1206 03:36:33.151411 4980 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Dec 06 03:36:33 crc kubenswrapper[4980]: I1206 03:36:33.174332 4980 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Dec 06 03:36:33 crc kubenswrapper[4980]: I1206 03:36:33.189219 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-p4gmt" Dec 06 03:36:33 crc kubenswrapper[4980]: I1206 03:36:33.190795 4980 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Dec 06 03:36:33 crc kubenswrapper[4980]: I1206 03:36:33.206019 4980 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-6txpx"] Dec 06 03:36:33 crc kubenswrapper[4980]: I1206 03:36:33.209579 4980 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 06 03:36:33 crc kubenswrapper[4980]: I1206 03:36:33.229625 4980 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 06 03:36:33 crc kubenswrapper[4980]: I1206 03:36:33.249833 4980 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Dec 06 03:36:33 crc kubenswrapper[4980]: I1206 03:36:33.324308 4980 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Dec 06 03:36:33 crc kubenswrapper[4980]: I1206 03:36:33.324501 4980 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Dec 06 03:36:33 crc kubenswrapper[4980]: I1206 03:36:33.324658 4980 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Dec 06 03:36:33 crc kubenswrapper[4980]: I1206 03:36:33.331355 4980 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Dec 06 03:36:33 crc kubenswrapper[4980]: I1206 03:36:33.354874 4980 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Dec 06 03:36:33 crc kubenswrapper[4980]: I1206 03:36:33.421456 4980 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Dec 06 03:36:33 crc kubenswrapper[4980]: I1206 03:36:33.421803 4980 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Dec 06 03:36:33 crc kubenswrapper[4980]: I1206 03:36:33.422007 4980 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Dec 06 03:36:33 crc kubenswrapper[4980]: I1206 03:36:33.446922 4980 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Dec 06 03:36:33 crc kubenswrapper[4980]: I1206 03:36:33.449401 4980 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Dec 06 03:36:33 crc kubenswrapper[4980]: I1206 03:36:33.468636 4980 request.go:700] Waited for 1.859059001s due to client-side throttling, not priority and fairness, request: GET:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-ingress-canary/configmaps?fieldSelector=metadata.name%3Dkube-root-ca.crt&limit=500&resourceVersion=0 Dec 06 03:36:33 crc kubenswrapper[4980]: I1206 03:36:33.470257 4980 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Dec 06 03:36:33 crc kubenswrapper[4980]: I1206 03:36:33.503895 4980 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-gjmj8"] Dec 06 03:36:33 crc kubenswrapper[4980]: I1206 03:36:33.508325 4980 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Dec 06 03:36:33 crc kubenswrapper[4980]: I1206 03:36:33.511421 4980 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Dec 06 03:36:33 crc kubenswrapper[4980]: I1206 03:36:33.529849 4980 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Dec 06 03:36:33 crc kubenswrapper[4980]: I1206 03:36:33.553542 4980 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Dec 06 03:36:33 crc kubenswrapper[4980]: I1206 03:36:33.575883 4980 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Dec 06 03:36:33 crc kubenswrapper[4980]: I1206 03:36:33.576102 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-425jl" event={"ID":"005c8e70-2c2b-43a4-ae31-fd9013840824","Type":"ContainerStarted","Data":"a084177d20462002e1c4464edacf4bfccb51ee03a14d84034eadb7e191e5b7dc"} Dec 06 03:36:33 crc kubenswrapper[4980]: I1206 03:36:33.576219 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-425jl" event={"ID":"005c8e70-2c2b-43a4-ae31-fd9013840824","Type":"ContainerStarted","Data":"b3925afb86396db6f4742faa911c284373ce11e138184626ed0e5638c5fc8242"} Dec 06 03:36:33 crc kubenswrapper[4980]: I1206 03:36:33.594241 4980 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Dec 06 03:36:33 crc kubenswrapper[4980]: I1206 03:36:33.602614 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-6txpx" event={"ID":"f2ea7661-b67f-4fbb-924a-943fe48995cb","Type":"ContainerStarted","Data":"1f89b147306bad18124152ade03af488c449385e51abb5cd45ad167f60df80b6"} Dec 06 03:36:33 crc kubenswrapper[4980]: I1206 03:36:33.610137 4980 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Dec 06 03:36:33 crc kubenswrapper[4980]: W1206 03:36:33.634371 4980 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod443a2dc9_0546_4ab0_a448_eab81c09b482.slice/crio-910dc6874f9328f4b53503bac76edcaf8fb09966143b238c1d0811e992afbc15 WatchSource:0}: Error finding container 910dc6874f9328f4b53503bac76edcaf8fb09966143b238c1d0811e992afbc15: Status 404 returned error can't find the container with id 910dc6874f9328f4b53503bac76edcaf8fb09966143b238c1d0811e992afbc15 Dec 06 03:36:33 crc kubenswrapper[4980]: I1206 03:36:33.654199 4980 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Dec 06 03:36:33 crc kubenswrapper[4980]: I1206 03:36:33.665793 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/2cf31abb-ef18-4ba5-8cc9-bf3f7507f771-image-import-ca\") pod \"apiserver-76f77b778f-s6h9g\" (UID: \"2cf31abb-ef18-4ba5-8cc9-bf3f7507f771\") " pod="openshift-apiserver/apiserver-76f77b778f-s6h9g" Dec 06 03:36:33 crc kubenswrapper[4980]: I1206 03:36:33.673236 4980 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Dec 06 03:36:33 crc kubenswrapper[4980]: I1206 03:36:33.685882 4980 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-2ph5v"] Dec 06 03:36:33 crc kubenswrapper[4980]: I1206 03:36:33.730914 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/768da7ce-a217-4749-be5b-c1ae2dd4e275-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-lrnhm\" (UID: \"768da7ce-a217-4749-be5b-c1ae2dd4e275\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-lrnhm" Dec 06 03:36:33 crc kubenswrapper[4980]: I1206 03:36:33.730960 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a3673c8a-c303-42ac-8565-f03e5d457aa1-serving-cert\") pod \"authentication-operator-69f744f599-84ft2\" (UID: \"a3673c8a-c303-42ac-8565-f03e5d457aa1\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-84ft2" Dec 06 03:36:33 crc kubenswrapper[4980]: I1206 03:36:33.731000 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/b25ba819-6e4e-4ca0-a570-c2726b8fbec5-ca-trust-extracted\") pod \"image-registry-697d97f7c8-qw2fb\" (UID: \"b25ba819-6e4e-4ca0-a570-c2726b8fbec5\") " pod="openshift-image-registry/image-registry-697d97f7c8-qw2fb" Dec 06 03:36:33 crc kubenswrapper[4980]: I1206 03:36:33.731025 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b25ba819-6e4e-4ca0-a570-c2726b8fbec5-trusted-ca\") pod \"image-registry-697d97f7c8-qw2fb\" (UID: \"b25ba819-6e4e-4ca0-a570-c2726b8fbec5\") " pod="openshift-image-registry/image-registry-697d97f7c8-qw2fb" Dec 06 03:36:33 crc kubenswrapper[4980]: I1206 03:36:33.731050 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kzm64\" (UniqueName: \"kubernetes.io/projected/a42aad68-42cf-4264-9c94-8678b0eea1b8-kube-api-access-kzm64\") pod \"oauth-openshift-558db77b4-pmmcw\" (UID: \"a42aad68-42cf-4264-9c94-8678b0eea1b8\") " pod="openshift-authentication/oauth-openshift-558db77b4-pmmcw" Dec 06 03:36:33 crc kubenswrapper[4980]: I1206 03:36:33.731070 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/dae23bd6-4c2b-441e-b44b-b6977bbd974b-metrics-tls\") pod \"dns-operator-744455d44c-w6bmg\" (UID: \"dae23bd6-4c2b-441e-b44b-b6977bbd974b\") " pod="openshift-dns-operator/dns-operator-744455d44c-w6bmg" Dec 06 03:36:33 crc kubenswrapper[4980]: I1206 03:36:33.731094 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/b25ba819-6e4e-4ca0-a570-c2726b8fbec5-installation-pull-secrets\") pod \"image-registry-697d97f7c8-qw2fb\" (UID: \"b25ba819-6e4e-4ca0-a570-c2726b8fbec5\") " pod="openshift-image-registry/image-registry-697d97f7c8-qw2fb" Dec 06 03:36:33 crc kubenswrapper[4980]: I1206 03:36:33.731127 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/b25ba819-6e4e-4ca0-a570-c2726b8fbec5-registry-tls\") pod \"image-registry-697d97f7c8-qw2fb\" (UID: \"b25ba819-6e4e-4ca0-a570-c2726b8fbec5\") " pod="openshift-image-registry/image-registry-697d97f7c8-qw2fb" Dec 06 03:36:33 crc kubenswrapper[4980]: I1206 03:36:33.731145 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/b25ba819-6e4e-4ca0-a570-c2726b8fbec5-registry-certificates\") pod \"image-registry-697d97f7c8-qw2fb\" (UID: \"b25ba819-6e4e-4ca0-a570-c2726b8fbec5\") " pod="openshift-image-registry/image-registry-697d97f7c8-qw2fb" Dec 06 03:36:33 crc kubenswrapper[4980]: I1206 03:36:33.731165 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bw764\" (UniqueName: \"kubernetes.io/projected/768da7ce-a217-4749-be5b-c1ae2dd4e275-kube-api-access-bw764\") pod \"openshift-controller-manager-operator-756b6f6bc6-lrnhm\" (UID: \"768da7ce-a217-4749-be5b-c1ae2dd4e275\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-lrnhm" Dec 06 03:36:33 crc kubenswrapper[4980]: I1206 03:36:33.731187 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/a42aad68-42cf-4264-9c94-8678b0eea1b8-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-pmmcw\" (UID: \"a42aad68-42cf-4264-9c94-8678b0eea1b8\") " pod="openshift-authentication/oauth-openshift-558db77b4-pmmcw" Dec 06 03:36:33 crc kubenswrapper[4980]: I1206 03:36:33.731210 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/a42aad68-42cf-4264-9c94-8678b0eea1b8-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-pmmcw\" (UID: \"a42aad68-42cf-4264-9c94-8678b0eea1b8\") " pod="openshift-authentication/oauth-openshift-558db77b4-pmmcw" Dec 06 03:36:33 crc kubenswrapper[4980]: I1206 03:36:33.731230 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a3673c8a-c303-42ac-8565-f03e5d457aa1-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-84ft2\" (UID: \"a3673c8a-c303-42ac-8565-f03e5d457aa1\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-84ft2" Dec 06 03:36:33 crc kubenswrapper[4980]: I1206 03:36:33.731249 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/a42aad68-42cf-4264-9c94-8678b0eea1b8-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-pmmcw\" (UID: \"a42aad68-42cf-4264-9c94-8678b0eea1b8\") " pod="openshift-authentication/oauth-openshift-558db77b4-pmmcw" Dec 06 03:36:33 crc kubenswrapper[4980]: I1206 03:36:33.731276 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qw2fb\" (UID: \"b25ba819-6e4e-4ca0-a570-c2726b8fbec5\") " pod="openshift-image-registry/image-registry-697d97f7c8-qw2fb" Dec 06 03:36:33 crc kubenswrapper[4980]: I1206 03:36:33.731294 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q2hf9\" (UniqueName: \"kubernetes.io/projected/6d695618-ca31-48d0-a178-24d7ee34a54b-kube-api-access-q2hf9\") pod \"console-operator-58897d9998-kftgf\" (UID: \"6d695618-ca31-48d0-a178-24d7ee34a54b\") " pod="openshift-console-operator/console-operator-58897d9998-kftgf" Dec 06 03:36:33 crc kubenswrapper[4980]: I1206 03:36:33.731328 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/768da7ce-a217-4749-be5b-c1ae2dd4e275-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-lrnhm\" (UID: \"768da7ce-a217-4749-be5b-c1ae2dd4e275\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-lrnhm" Dec 06 03:36:33 crc kubenswrapper[4980]: I1206 03:36:33.731345 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/a42aad68-42cf-4264-9c94-8678b0eea1b8-audit-dir\") pod \"oauth-openshift-558db77b4-pmmcw\" (UID: \"a42aad68-42cf-4264-9c94-8678b0eea1b8\") " pod="openshift-authentication/oauth-openshift-558db77b4-pmmcw" Dec 06 03:36:33 crc kubenswrapper[4980]: I1206 03:36:33.731366 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/a42aad68-42cf-4264-9c94-8678b0eea1b8-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-pmmcw\" (UID: \"a42aad68-42cf-4264-9c94-8678b0eea1b8\") " pod="openshift-authentication/oauth-openshift-558db77b4-pmmcw" Dec 06 03:36:33 crc kubenswrapper[4980]: I1206 03:36:33.731383 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a42aad68-42cf-4264-9c94-8678b0eea1b8-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-pmmcw\" (UID: \"a42aad68-42cf-4264-9c94-8678b0eea1b8\") " pod="openshift-authentication/oauth-openshift-558db77b4-pmmcw" Dec 06 03:36:33 crc kubenswrapper[4980]: I1206 03:36:33.731411 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/a42aad68-42cf-4264-9c94-8678b0eea1b8-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-pmmcw\" (UID: \"a42aad68-42cf-4264-9c94-8678b0eea1b8\") " pod="openshift-authentication/oauth-openshift-558db77b4-pmmcw" Dec 06 03:36:33 crc kubenswrapper[4980]: I1206 03:36:33.731431 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bq8xx\" (UniqueName: \"kubernetes.io/projected/a3673c8a-c303-42ac-8565-f03e5d457aa1-kube-api-access-bq8xx\") pod \"authentication-operator-69f744f599-84ft2\" (UID: \"a3673c8a-c303-42ac-8565-f03e5d457aa1\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-84ft2" Dec 06 03:36:33 crc kubenswrapper[4980]: I1206 03:36:33.731447 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/6d695618-ca31-48d0-a178-24d7ee34a54b-trusted-ca\") pod \"console-operator-58897d9998-kftgf\" (UID: \"6d695618-ca31-48d0-a178-24d7ee34a54b\") " pod="openshift-console-operator/console-operator-58897d9998-kftgf" Dec 06 03:36:33 crc kubenswrapper[4980]: I1206 03:36:33.731475 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/a42aad68-42cf-4264-9c94-8678b0eea1b8-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-pmmcw\" (UID: \"a42aad68-42cf-4264-9c94-8678b0eea1b8\") " pod="openshift-authentication/oauth-openshift-558db77b4-pmmcw" Dec 06 03:36:33 crc kubenswrapper[4980]: I1206 03:36:33.731491 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/b25ba819-6e4e-4ca0-a570-c2726b8fbec5-bound-sa-token\") pod \"image-registry-697d97f7c8-qw2fb\" (UID: \"b25ba819-6e4e-4ca0-a570-c2726b8fbec5\") " pod="openshift-image-registry/image-registry-697d97f7c8-qw2fb" Dec 06 03:36:33 crc kubenswrapper[4980]: I1206 03:36:33.731505 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bzls8\" (UniqueName: \"kubernetes.io/projected/b25ba819-6e4e-4ca0-a570-c2726b8fbec5-kube-api-access-bzls8\") pod \"image-registry-697d97f7c8-qw2fb\" (UID: \"b25ba819-6e4e-4ca0-a570-c2726b8fbec5\") " pod="openshift-image-registry/image-registry-697d97f7c8-qw2fb" Dec 06 03:36:33 crc kubenswrapper[4980]: I1206 03:36:33.731552 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/a42aad68-42cf-4264-9c94-8678b0eea1b8-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-pmmcw\" (UID: \"a42aad68-42cf-4264-9c94-8678b0eea1b8\") " pod="openshift-authentication/oauth-openshift-558db77b4-pmmcw" Dec 06 03:36:33 crc kubenswrapper[4980]: I1206 03:36:33.731576 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6d695618-ca31-48d0-a178-24d7ee34a54b-config\") pod \"console-operator-58897d9998-kftgf\" (UID: \"6d695618-ca31-48d0-a178-24d7ee34a54b\") " pod="openshift-console-operator/console-operator-58897d9998-kftgf" Dec 06 03:36:33 crc kubenswrapper[4980]: I1206 03:36:33.731595 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a3673c8a-c303-42ac-8565-f03e5d457aa1-service-ca-bundle\") pod \"authentication-operator-69f744f599-84ft2\" (UID: \"a3673c8a-c303-42ac-8565-f03e5d457aa1\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-84ft2" Dec 06 03:36:33 crc kubenswrapper[4980]: I1206 03:36:33.732133 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a3673c8a-c303-42ac-8565-f03e5d457aa1-config\") pod \"authentication-operator-69f744f599-84ft2\" (UID: \"a3673c8a-c303-42ac-8565-f03e5d457aa1\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-84ft2" Dec 06 03:36:33 crc kubenswrapper[4980]: I1206 03:36:33.732193 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8kssv\" (UniqueName: \"kubernetes.io/projected/346bb011-c34e-4286-90d3-0242205e8e94-kube-api-access-8kssv\") pod \"downloads-7954f5f757-6h96g\" (UID: \"346bb011-c34e-4286-90d3-0242205e8e94\") " pod="openshift-console/downloads-7954f5f757-6h96g" Dec 06 03:36:33 crc kubenswrapper[4980]: I1206 03:36:33.732277 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/a42aad68-42cf-4264-9c94-8678b0eea1b8-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-pmmcw\" (UID: \"a42aad68-42cf-4264-9c94-8678b0eea1b8\") " pod="openshift-authentication/oauth-openshift-558db77b4-pmmcw" Dec 06 03:36:33 crc kubenswrapper[4980]: I1206 03:36:33.732327 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/a42aad68-42cf-4264-9c94-8678b0eea1b8-audit-policies\") pod \"oauth-openshift-558db77b4-pmmcw\" (UID: \"a42aad68-42cf-4264-9c94-8678b0eea1b8\") " pod="openshift-authentication/oauth-openshift-558db77b4-pmmcw" Dec 06 03:36:33 crc kubenswrapper[4980]: I1206 03:36:33.732348 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/a42aad68-42cf-4264-9c94-8678b0eea1b8-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-pmmcw\" (UID: \"a42aad68-42cf-4264-9c94-8678b0eea1b8\") " pod="openshift-authentication/oauth-openshift-558db77b4-pmmcw" Dec 06 03:36:33 crc kubenswrapper[4980]: I1206 03:36:33.732367 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/a42aad68-42cf-4264-9c94-8678b0eea1b8-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-pmmcw\" (UID: \"a42aad68-42cf-4264-9c94-8678b0eea1b8\") " pod="openshift-authentication/oauth-openshift-558db77b4-pmmcw" Dec 06 03:36:33 crc kubenswrapper[4980]: I1206 03:36:33.732387 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6d695618-ca31-48d0-a178-24d7ee34a54b-serving-cert\") pod \"console-operator-58897d9998-kftgf\" (UID: \"6d695618-ca31-48d0-a178-24d7ee34a54b\") " pod="openshift-console-operator/console-operator-58897d9998-kftgf" Dec 06 03:36:33 crc kubenswrapper[4980]: I1206 03:36:33.732416 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-257vm\" (UniqueName: \"kubernetes.io/projected/dae23bd6-4c2b-441e-b44b-b6977bbd974b-kube-api-access-257vm\") pod \"dns-operator-744455d44c-w6bmg\" (UID: \"dae23bd6-4c2b-441e-b44b-b6977bbd974b\") " pod="openshift-dns-operator/dns-operator-744455d44c-w6bmg" Dec 06 03:36:33 crc kubenswrapper[4980]: E1206 03:36:33.732505 4980 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 03:36:34.232480832 +0000 UTC m=+153.477232283 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qw2fb" (UID: "b25ba819-6e4e-4ca0-a570-c2726b8fbec5") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 03:36:33 crc kubenswrapper[4980]: I1206 03:36:33.734541 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/2cf31abb-ef18-4ba5-8cc9-bf3f7507f771-etcd-client\") pod \"apiserver-76f77b778f-s6h9g\" (UID: \"2cf31abb-ef18-4ba5-8cc9-bf3f7507f771\") " pod="openshift-apiserver/apiserver-76f77b778f-s6h9g" Dec 06 03:36:33 crc kubenswrapper[4980]: I1206 03:36:33.860457 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 03:36:33 crc kubenswrapper[4980]: I1206 03:36:33.860611 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/a42aad68-42cf-4264-9c94-8678b0eea1b8-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-pmmcw\" (UID: \"a42aad68-42cf-4264-9c94-8678b0eea1b8\") " pod="openshift-authentication/oauth-openshift-558db77b4-pmmcw" Dec 06 03:36:33 crc kubenswrapper[4980]: I1206 03:36:33.860651 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a3673c8a-c303-42ac-8565-f03e5d457aa1-service-ca-bundle\") pod \"authentication-operator-69f744f599-84ft2\" (UID: \"a3673c8a-c303-42ac-8565-f03e5d457aa1\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-84ft2" Dec 06 03:36:33 crc kubenswrapper[4980]: I1206 03:36:33.860668 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/b25ba819-6e4e-4ca0-a570-c2726b8fbec5-bound-sa-token\") pod \"image-registry-697d97f7c8-qw2fb\" (UID: \"b25ba819-6e4e-4ca0-a570-c2726b8fbec5\") " pod="openshift-image-registry/image-registry-697d97f7c8-qw2fb" Dec 06 03:36:33 crc kubenswrapper[4980]: I1206 03:36:33.860686 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bzls8\" (UniqueName: \"kubernetes.io/projected/b25ba819-6e4e-4ca0-a570-c2726b8fbec5-kube-api-access-bzls8\") pod \"image-registry-697d97f7c8-qw2fb\" (UID: \"b25ba819-6e4e-4ca0-a570-c2726b8fbec5\") " pod="openshift-image-registry/image-registry-697d97f7c8-qw2fb" Dec 06 03:36:33 crc kubenswrapper[4980]: I1206 03:36:33.860706 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mfkfg\" (UniqueName: \"kubernetes.io/projected/d7fb8e4e-88e7-4cc9-baf9-6e217813b28d-kube-api-access-mfkfg\") pod \"migrator-59844c95c7-56s8p\" (UID: \"d7fb8e4e-88e7-4cc9-baf9-6e217813b28d\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-56s8p" Dec 06 03:36:33 crc kubenswrapper[4980]: I1206 03:36:33.860724 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/acb53ac6-3f09-424c-a830-a1e15eec0d62-service-ca-bundle\") pod \"router-default-5444994796-pzzkd\" (UID: \"acb53ac6-3f09-424c-a830-a1e15eec0d62\") " pod="openshift-ingress/router-default-5444994796-pzzkd" Dec 06 03:36:33 crc kubenswrapper[4980]: I1206 03:36:33.860749 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/f08eff3e-89fe-4caf-9ec1-b32ee99f3c10-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-xjtnv\" (UID: \"f08eff3e-89fe-4caf-9ec1-b32ee99f3c10\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-xjtnv" Dec 06 03:36:33 crc kubenswrapper[4980]: I1206 03:36:33.860811 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dp275\" (UniqueName: \"kubernetes.io/projected/ccd9c9f8-3473-4f3d-884c-e545a0a502af-kube-api-access-dp275\") pod \"kube-storage-version-migrator-operator-b67b599dd-cjvdn\" (UID: \"ccd9c9f8-3473-4f3d-884c-e545a0a502af\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-cjvdn" Dec 06 03:36:33 crc kubenswrapper[4980]: I1206 03:36:33.860829 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t52g4\" (UniqueName: \"kubernetes.io/projected/3ee9a5b6-d84f-46d7-b3c2-aab85c634bd4-kube-api-access-t52g4\") pod \"ingress-canary-799j5\" (UID: \"3ee9a5b6-d84f-46d7-b3c2-aab85c634bd4\") " pod="openshift-ingress-canary/ingress-canary-799j5" Dec 06 03:36:33 crc kubenswrapper[4980]: I1206 03:36:33.860880 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/46290605-e114-4beb-b1c9-b0389f20f5f3-bound-sa-token\") pod \"ingress-operator-5b745b69d9-27gc7\" (UID: \"46290605-e114-4beb-b1c9-b0389f20f5f3\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-27gc7" Dec 06 03:36:33 crc kubenswrapper[4980]: I1206 03:36:33.860905 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8kssv\" (UniqueName: \"kubernetes.io/projected/346bb011-c34e-4286-90d3-0242205e8e94-kube-api-access-8kssv\") pod \"downloads-7954f5f757-6h96g\" (UID: \"346bb011-c34e-4286-90d3-0242205e8e94\") " pod="openshift-console/downloads-7954f5f757-6h96g" Dec 06 03:36:33 crc kubenswrapper[4980]: I1206 03:36:33.860921 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a3673c8a-c303-42ac-8565-f03e5d457aa1-config\") pod \"authentication-operator-69f744f599-84ft2\" (UID: \"a3673c8a-c303-42ac-8565-f03e5d457aa1\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-84ft2" Dec 06 03:36:33 crc kubenswrapper[4980]: I1206 03:36:33.860962 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/92b5e190-2712-40a7-a551-472805f3c84e-certs\") pod \"machine-config-server-bfp7v\" (UID: \"92b5e190-2712-40a7-a551-472805f3c84e\") " pod="openshift-machine-config-operator/machine-config-server-bfp7v" Dec 06 03:36:33 crc kubenswrapper[4980]: I1206 03:36:33.860985 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/a42aad68-42cf-4264-9c94-8678b0eea1b8-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-pmmcw\" (UID: \"a42aad68-42cf-4264-9c94-8678b0eea1b8\") " pod="openshift-authentication/oauth-openshift-558db77b4-pmmcw" Dec 06 03:36:33 crc kubenswrapper[4980]: I1206 03:36:33.861004 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/92b5e190-2712-40a7-a551-472805f3c84e-node-bootstrap-token\") pod \"machine-config-server-bfp7v\" (UID: \"92b5e190-2712-40a7-a551-472805f3c84e\") " pod="openshift-machine-config-operator/machine-config-server-bfp7v" Dec 06 03:36:33 crc kubenswrapper[4980]: I1206 03:36:33.861040 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/0a45bc54-be9c-4c4f-9cae-dc369d25be33-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-spd2v\" (UID: \"0a45bc54-be9c-4c4f-9cae-dc369d25be33\") " pod="openshift-marketplace/marketplace-operator-79b997595-spd2v" Dec 06 03:36:33 crc kubenswrapper[4980]: I1206 03:36:33.861060 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6d695618-ca31-48d0-a178-24d7ee34a54b-serving-cert\") pod \"console-operator-58897d9998-kftgf\" (UID: \"6d695618-ca31-48d0-a178-24d7ee34a54b\") " pod="openshift-console-operator/console-operator-58897d9998-kftgf" Dec 06 03:36:33 crc kubenswrapper[4980]: I1206 03:36:33.861078 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wr5kj\" (UniqueName: \"kubernetes.io/projected/05aaf766-565d-4b05-97ec-6c8078d33f85-kube-api-access-wr5kj\") pod \"machine-config-operator-74547568cd-49749\" (UID: \"05aaf766-565d-4b05-97ec-6c8078d33f85\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-49749" Dec 06 03:36:33 crc kubenswrapper[4980]: I1206 03:36:33.861119 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/8be432d7-9e4b-4452-81df-91998326f92b-metrics-tls\") pod \"dns-default-4dcjh\" (UID: \"8be432d7-9e4b-4452-81df-91998326f92b\") " pod="openshift-dns/dns-default-4dcjh" Dec 06 03:36:33 crc kubenswrapper[4980]: I1206 03:36:33.861138 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/a42aad68-42cf-4264-9c94-8678b0eea1b8-audit-policies\") pod \"oauth-openshift-558db77b4-pmmcw\" (UID: \"a42aad68-42cf-4264-9c94-8678b0eea1b8\") " pod="openshift-authentication/oauth-openshift-558db77b4-pmmcw" Dec 06 03:36:33 crc kubenswrapper[4980]: I1206 03:36:33.861155 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/a42aad68-42cf-4264-9c94-8678b0eea1b8-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-pmmcw\" (UID: \"a42aad68-42cf-4264-9c94-8678b0eea1b8\") " pod="openshift-authentication/oauth-openshift-558db77b4-pmmcw" Dec 06 03:36:33 crc kubenswrapper[4980]: I1206 03:36:33.861175 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/0a45bc54-be9c-4c4f-9cae-dc369d25be33-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-spd2v\" (UID: \"0a45bc54-be9c-4c4f-9cae-dc369d25be33\") " pod="openshift-marketplace/marketplace-operator-79b997595-spd2v" Dec 06 03:36:33 crc kubenswrapper[4980]: I1206 03:36:33.861220 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ccd9c9f8-3473-4f3d-884c-e545a0a502af-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-cjvdn\" (UID: \"ccd9c9f8-3473-4f3d-884c-e545a0a502af\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-cjvdn" Dec 06 03:36:33 crc kubenswrapper[4980]: I1206 03:36:33.861236 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b2da6efd-2b6d-4ca7-b056-32c4b132becf-config\") pod \"service-ca-operator-777779d784-r47r5\" (UID: \"b2da6efd-2b6d-4ca7-b056-32c4b132becf\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-r47r5" Dec 06 03:36:33 crc kubenswrapper[4980]: I1206 03:36:33.861293 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/05aaf766-565d-4b05-97ec-6c8078d33f85-auth-proxy-config\") pod \"machine-config-operator-74547568cd-49749\" (UID: \"05aaf766-565d-4b05-97ec-6c8078d33f85\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-49749" Dec 06 03:36:33 crc kubenswrapper[4980]: I1206 03:36:33.861328 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f08eff3e-89fe-4caf-9ec1-b32ee99f3c10-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-xjtnv\" (UID: \"f08eff3e-89fe-4caf-9ec1-b32ee99f3c10\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-xjtnv" Dec 06 03:36:33 crc kubenswrapper[4980]: I1206 03:36:33.861371 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vcwpf\" (UniqueName: \"kubernetes.io/projected/9476a6a0-4d40-4c85-9bd1-2162c971e3dd-kube-api-access-vcwpf\") pod \"olm-operator-6b444d44fb-sq6bp\" (UID: \"9476a6a0-4d40-4c85-9bd1-2162c971e3dd\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-sq6bp" Dec 06 03:36:33 crc kubenswrapper[4980]: I1206 03:36:33.861405 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t26fr\" (UniqueName: \"kubernetes.io/projected/b2da6efd-2b6d-4ca7-b056-32c4b132becf-kube-api-access-t26fr\") pod \"service-ca-operator-777779d784-r47r5\" (UID: \"b2da6efd-2b6d-4ca7-b056-32c4b132becf\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-r47r5" Dec 06 03:36:33 crc kubenswrapper[4980]: I1206 03:36:33.861464 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/9476a6a0-4d40-4c85-9bd1-2162c971e3dd-srv-cert\") pod \"olm-operator-6b444d44fb-sq6bp\" (UID: \"9476a6a0-4d40-4c85-9bd1-2162c971e3dd\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-sq6bp" Dec 06 03:36:33 crc kubenswrapper[4980]: I1206 03:36:33.861504 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/768da7ce-a217-4749-be5b-c1ae2dd4e275-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-lrnhm\" (UID: \"768da7ce-a217-4749-be5b-c1ae2dd4e275\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-lrnhm" Dec 06 03:36:33 crc kubenswrapper[4980]: I1206 03:36:33.861549 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a3673c8a-c303-42ac-8565-f03e5d457aa1-serving-cert\") pod \"authentication-operator-69f744f599-84ft2\" (UID: \"a3673c8a-c303-42ac-8565-f03e5d457aa1\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-84ft2" Dec 06 03:36:33 crc kubenswrapper[4980]: I1206 03:36:33.861631 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/b25ba819-6e4e-4ca0-a570-c2726b8fbec5-ca-trust-extracted\") pod \"image-registry-697d97f7c8-qw2fb\" (UID: \"b25ba819-6e4e-4ca0-a570-c2726b8fbec5\") " pod="openshift-image-registry/image-registry-697d97f7c8-qw2fb" Dec 06 03:36:33 crc kubenswrapper[4980]: I1206 03:36:33.861649 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/46290605-e114-4beb-b1c9-b0389f20f5f3-metrics-tls\") pod \"ingress-operator-5b745b69d9-27gc7\" (UID: \"46290605-e114-4beb-b1c9-b0389f20f5f3\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-27gc7" Dec 06 03:36:33 crc kubenswrapper[4980]: I1206 03:36:33.861665 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/46290605-e114-4beb-b1c9-b0389f20f5f3-trusted-ca\") pod \"ingress-operator-5b745b69d9-27gc7\" (UID: \"46290605-e114-4beb-b1c9-b0389f20f5f3\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-27gc7" Dec 06 03:36:33 crc kubenswrapper[4980]: I1206 03:36:33.861725 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/b25ba819-6e4e-4ca0-a570-c2726b8fbec5-installation-pull-secrets\") pod \"image-registry-697d97f7c8-qw2fb\" (UID: \"b25ba819-6e4e-4ca0-a570-c2726b8fbec5\") " pod="openshift-image-registry/image-registry-697d97f7c8-qw2fb" Dec 06 03:36:33 crc kubenswrapper[4980]: I1206 03:36:33.861801 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kzm64\" (UniqueName: \"kubernetes.io/projected/a42aad68-42cf-4264-9c94-8678b0eea1b8-kube-api-access-kzm64\") pod \"oauth-openshift-558db77b4-pmmcw\" (UID: \"a42aad68-42cf-4264-9c94-8678b0eea1b8\") " pod="openshift-authentication/oauth-openshift-558db77b4-pmmcw" Dec 06 03:36:33 crc kubenswrapper[4980]: I1206 03:36:33.861852 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/dae23bd6-4c2b-441e-b44b-b6977bbd974b-metrics-tls\") pod \"dns-operator-744455d44c-w6bmg\" (UID: \"dae23bd6-4c2b-441e-b44b-b6977bbd974b\") " pod="openshift-dns-operator/dns-operator-744455d44c-w6bmg" Dec 06 03:36:33 crc kubenswrapper[4980]: I1206 03:36:33.861897 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z4m5n\" (UniqueName: \"kubernetes.io/projected/abfe3c42-9229-486f-8c87-8b883f6821e0-kube-api-access-z4m5n\") pod \"machine-config-controller-84d6567774-d6dd4\" (UID: \"abfe3c42-9229-486f-8c87-8b883f6821e0\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-d6dd4" Dec 06 03:36:33 crc kubenswrapper[4980]: I1206 03:36:33.861920 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/e81d9a6d-991f-46b0-aba5-6f76a5b1031a-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-cqk6j\" (UID: \"e81d9a6d-991f-46b0-aba5-6f76a5b1031a\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-cqk6j" Dec 06 03:36:33 crc kubenswrapper[4980]: I1206 03:36:33.861953 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/a42aad68-42cf-4264-9c94-8678b0eea1b8-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-pmmcw\" (UID: \"a42aad68-42cf-4264-9c94-8678b0eea1b8\") " pod="openshift-authentication/oauth-openshift-558db77b4-pmmcw" Dec 06 03:36:33 crc kubenswrapper[4980]: I1206 03:36:33.861971 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/03704eb9-227d-4985-87cf-acd66ae76037-config-volume\") pod \"collect-profiles-29416530-9hm9m\" (UID: \"03704eb9-227d-4985-87cf-acd66ae76037\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416530-9hm9m" Dec 06 03:36:33 crc kubenswrapper[4980]: I1206 03:36:33.861992 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/e1e4e0b4-0bfe-4b01-9c4e-710ba0ec03bd-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-qcn9f\" (UID: \"e1e4e0b4-0bfe-4b01-9c4e-710ba0ec03bd\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-qcn9f" Dec 06 03:36:33 crc kubenswrapper[4980]: I1206 03:36:33.862012 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/b25ba819-6e4e-4ca0-a570-c2726b8fbec5-registry-tls\") pod \"image-registry-697d97f7c8-qw2fb\" (UID: \"b25ba819-6e4e-4ca0-a570-c2726b8fbec5\") " pod="openshift-image-registry/image-registry-697d97f7c8-qw2fb" Dec 06 03:36:33 crc kubenswrapper[4980]: I1206 03:36:33.862034 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/b25ba819-6e4e-4ca0-a570-c2726b8fbec5-registry-certificates\") pod \"image-registry-697d97f7c8-qw2fb\" (UID: \"b25ba819-6e4e-4ca0-a570-c2726b8fbec5\") " pod="openshift-image-registry/image-registry-697d97f7c8-qw2fb" Dec 06 03:36:33 crc kubenswrapper[4980]: I1206 03:36:33.862053 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bw764\" (UniqueName: \"kubernetes.io/projected/768da7ce-a217-4749-be5b-c1ae2dd4e275-kube-api-access-bw764\") pod \"openshift-controller-manager-operator-756b6f6bc6-lrnhm\" (UID: \"768da7ce-a217-4749-be5b-c1ae2dd4e275\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-lrnhm" Dec 06 03:36:33 crc kubenswrapper[4980]: I1206 03:36:33.862086 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/a42aad68-42cf-4264-9c94-8678b0eea1b8-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-pmmcw\" (UID: \"a42aad68-42cf-4264-9c94-8678b0eea1b8\") " pod="openshift-authentication/oauth-openshift-558db77b4-pmmcw" Dec 06 03:36:33 crc kubenswrapper[4980]: I1206 03:36:33.862300 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/abfe3c42-9229-486f-8c87-8b883f6821e0-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-d6dd4\" (UID: \"abfe3c42-9229-486f-8c87-8b883f6821e0\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-d6dd4" Dec 06 03:36:33 crc kubenswrapper[4980]: I1206 03:36:33.862329 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f2555ff9-efdb-449a-9f88-67a3c80ad0cf-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-jp25g\" (UID: \"f2555ff9-efdb-449a-9f88-67a3c80ad0cf\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-jp25g" Dec 06 03:36:33 crc kubenswrapper[4980]: I1206 03:36:33.862357 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9p5zn\" (UniqueName: \"kubernetes.io/projected/46290605-e114-4beb-b1c9-b0389f20f5f3-kube-api-access-9p5zn\") pod \"ingress-operator-5b745b69d9-27gc7\" (UID: \"46290605-e114-4beb-b1c9-b0389f20f5f3\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-27gc7" Dec 06 03:36:33 crc kubenswrapper[4980]: I1206 03:36:33.862402 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x6x86\" (UniqueName: \"kubernetes.io/projected/a7ee597a-a4fc-45ef-841f-1d4f666ccfcf-kube-api-access-x6x86\") pod \"catalog-operator-68c6474976-hjrkf\" (UID: \"a7ee597a-a4fc-45ef-841f-1d4f666ccfcf\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-hjrkf" Dec 06 03:36:33 crc kubenswrapper[4980]: I1206 03:36:33.862434 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/a42aad68-42cf-4264-9c94-8678b0eea1b8-audit-dir\") pod \"oauth-openshift-558db77b4-pmmcw\" (UID: \"a42aad68-42cf-4264-9c94-8678b0eea1b8\") " pod="openshift-authentication/oauth-openshift-558db77b4-pmmcw" Dec 06 03:36:33 crc kubenswrapper[4980]: I1206 03:36:33.862476 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/a42aad68-42cf-4264-9c94-8678b0eea1b8-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-pmmcw\" (UID: \"a42aad68-42cf-4264-9c94-8678b0eea1b8\") " pod="openshift-authentication/oauth-openshift-558db77b4-pmmcw" Dec 06 03:36:33 crc kubenswrapper[4980]: I1206 03:36:33.862504 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/3ee9a5b6-d84f-46d7-b3c2-aab85c634bd4-cert\") pod \"ingress-canary-799j5\" (UID: \"3ee9a5b6-d84f-46d7-b3c2-aab85c634bd4\") " pod="openshift-ingress-canary/ingress-canary-799j5" Dec 06 03:36:33 crc kubenswrapper[4980]: I1206 03:36:33.862548 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/9b9bddd5-f192-4b4c-81ee-b523bf9c8705-plugins-dir\") pod \"csi-hostpathplugin-74hx2\" (UID: \"9b9bddd5-f192-4b4c-81ee-b523bf9c8705\") " pod="hostpath-provisioner/csi-hostpathplugin-74hx2" Dec 06 03:36:33 crc kubenswrapper[4980]: I1206 03:36:33.862575 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/a42aad68-42cf-4264-9c94-8678b0eea1b8-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-pmmcw\" (UID: \"a42aad68-42cf-4264-9c94-8678b0eea1b8\") " pod="openshift-authentication/oauth-openshift-558db77b4-pmmcw" Dec 06 03:36:33 crc kubenswrapper[4980]: I1206 03:36:33.862599 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/c7146eb0-acc5-4bac-b429-e1e00761e666-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-8jc8f\" (UID: \"c7146eb0-acc5-4bac-b429-e1e00761e666\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-8jc8f" Dec 06 03:36:33 crc kubenswrapper[4980]: E1206 03:36:33.862852 4980 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 03:36:34.362826915 +0000 UTC m=+153.607578186 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 03:36:33 crc kubenswrapper[4980]: I1206 03:36:33.863086 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bq8xx\" (UniqueName: \"kubernetes.io/projected/a3673c8a-c303-42ac-8565-f03e5d457aa1-kube-api-access-bq8xx\") pod \"authentication-operator-69f744f599-84ft2\" (UID: \"a3673c8a-c303-42ac-8565-f03e5d457aa1\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-84ft2" Dec 06 03:36:33 crc kubenswrapper[4980]: I1206 03:36:33.863111 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/2692cb89-853a-4c4e-a6a6-570673db6f14-signing-key\") pod \"service-ca-9c57cc56f-qpf6b\" (UID: \"2692cb89-853a-4c4e-a6a6-570673db6f14\") " pod="openshift-service-ca/service-ca-9c57cc56f-qpf6b" Dec 06 03:36:33 crc kubenswrapper[4980]: I1206 03:36:33.863150 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/9b9bddd5-f192-4b4c-81ee-b523bf9c8705-socket-dir\") pod \"csi-hostpathplugin-74hx2\" (UID: \"9b9bddd5-f192-4b4c-81ee-b523bf9c8705\") " pod="hostpath-provisioner/csi-hostpathplugin-74hx2" Dec 06 03:36:33 crc kubenswrapper[4980]: I1206 03:36:33.863171 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/9b9bddd5-f192-4b4c-81ee-b523bf9c8705-registration-dir\") pod \"csi-hostpathplugin-74hx2\" (UID: \"9b9bddd5-f192-4b4c-81ee-b523bf9c8705\") " pod="hostpath-provisioner/csi-hostpathplugin-74hx2" Dec 06 03:36:33 crc kubenswrapper[4980]: I1206 03:36:33.863225 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/6d695618-ca31-48d0-a178-24d7ee34a54b-trusted-ca\") pod \"console-operator-58897d9998-kftgf\" (UID: \"6d695618-ca31-48d0-a178-24d7ee34a54b\") " pod="openshift-console-operator/console-operator-58897d9998-kftgf" Dec 06 03:36:33 crc kubenswrapper[4980]: I1206 03:36:33.863272 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/949f499a-5680-4fec-ab59-5ced4f24ba2f-webhook-cert\") pod \"packageserver-d55dfcdfc-gjdsq\" (UID: \"949f499a-5680-4fec-ab59-5ced4f24ba2f\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-gjdsq" Dec 06 03:36:33 crc kubenswrapper[4980]: I1206 03:36:33.863312 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f2555ff9-efdb-449a-9f88-67a3c80ad0cf-config\") pod \"kube-controller-manager-operator-78b949d7b-jp25g\" (UID: \"f2555ff9-efdb-449a-9f88-67a3c80ad0cf\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-jp25g" Dec 06 03:36:33 crc kubenswrapper[4980]: I1206 03:36:33.863334 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vg8kf\" (UniqueName: \"kubernetes.io/projected/e81d9a6d-991f-46b0-aba5-6f76a5b1031a-kube-api-access-vg8kf\") pod \"control-plane-machine-set-operator-78cbb6b69f-cqk6j\" (UID: \"e81d9a6d-991f-46b0-aba5-6f76a5b1031a\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-cqk6j" Dec 06 03:36:33 crc kubenswrapper[4980]: I1206 03:36:33.863353 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f5l5z\" (UniqueName: \"kubernetes.io/projected/0a45bc54-be9c-4c4f-9cae-dc369d25be33-kube-api-access-f5l5z\") pod \"marketplace-operator-79b997595-spd2v\" (UID: \"0a45bc54-be9c-4c4f-9cae-dc369d25be33\") " pod="openshift-marketplace/marketplace-operator-79b997595-spd2v" Dec 06 03:36:33 crc kubenswrapper[4980]: I1206 03:36:33.863391 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/9b9bddd5-f192-4b4c-81ee-b523bf9c8705-csi-data-dir\") pod \"csi-hostpathplugin-74hx2\" (UID: \"9b9bddd5-f192-4b4c-81ee-b523bf9c8705\") " pod="hostpath-provisioner/csi-hostpathplugin-74hx2" Dec 06 03:36:33 crc kubenswrapper[4980]: I1206 03:36:33.863412 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6d695618-ca31-48d0-a178-24d7ee34a54b-config\") pod \"console-operator-58897d9998-kftgf\" (UID: \"6d695618-ca31-48d0-a178-24d7ee34a54b\") " pod="openshift-console-operator/console-operator-58897d9998-kftgf" Dec 06 03:36:33 crc kubenswrapper[4980]: I1206 03:36:33.863429 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/03704eb9-227d-4985-87cf-acd66ae76037-secret-volume\") pod \"collect-profiles-29416530-9hm9m\" (UID: \"03704eb9-227d-4985-87cf-acd66ae76037\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416530-9hm9m" Dec 06 03:36:33 crc kubenswrapper[4980]: I1206 03:36:33.863477 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/a7ee597a-a4fc-45ef-841f-1d4f666ccfcf-profile-collector-cert\") pod \"catalog-operator-68c6474976-hjrkf\" (UID: \"a7ee597a-a4fc-45ef-841f-1d4f666ccfcf\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-hjrkf" Dec 06 03:36:33 crc kubenswrapper[4980]: I1206 03:36:33.863541 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/f2555ff9-efdb-449a-9f88-67a3c80ad0cf-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-jp25g\" (UID: \"f2555ff9-efdb-449a-9f88-67a3c80ad0cf\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-jp25g" Dec 06 03:36:33 crc kubenswrapper[4980]: I1206 03:36:33.863564 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/a42aad68-42cf-4264-9c94-8678b0eea1b8-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-pmmcw\" (UID: \"a42aad68-42cf-4264-9c94-8678b0eea1b8\") " pod="openshift-authentication/oauth-openshift-558db77b4-pmmcw" Dec 06 03:36:33 crc kubenswrapper[4980]: I1206 03:36:33.863583 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-257vm\" (UniqueName: \"kubernetes.io/projected/dae23bd6-4c2b-441e-b44b-b6977bbd974b-kube-api-access-257vm\") pod \"dns-operator-744455d44c-w6bmg\" (UID: \"dae23bd6-4c2b-441e-b44b-b6977bbd974b\") " pod="openshift-dns-operator/dns-operator-744455d44c-w6bmg" Dec 06 03:36:33 crc kubenswrapper[4980]: I1206 03:36:33.863622 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b6q5n\" (UniqueName: \"kubernetes.io/projected/8be432d7-9e4b-4452-81df-91998326f92b-kube-api-access-b6q5n\") pod \"dns-default-4dcjh\" (UID: \"8be432d7-9e4b-4452-81df-91998326f92b\") " pod="openshift-dns/dns-default-4dcjh" Dec 06 03:36:33 crc kubenswrapper[4980]: I1206 03:36:33.863656 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/d1d006b0-92b0-45d5-9c28-fd43f879b310-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-tlpd8\" (UID: \"d1d006b0-92b0-45d5-9c28-fd43f879b310\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-tlpd8" Dec 06 03:36:33 crc kubenswrapper[4980]: I1206 03:36:33.863698 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2w6nc\" (UniqueName: \"kubernetes.io/projected/03704eb9-227d-4985-87cf-acd66ae76037-kube-api-access-2w6nc\") pod \"collect-profiles-29416530-9hm9m\" (UID: \"03704eb9-227d-4985-87cf-acd66ae76037\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416530-9hm9m" Dec 06 03:36:33 crc kubenswrapper[4980]: I1206 03:36:33.863720 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pl4qc\" (UniqueName: \"kubernetes.io/projected/9b9bddd5-f192-4b4c-81ee-b523bf9c8705-kube-api-access-pl4qc\") pod \"csi-hostpathplugin-74hx2\" (UID: \"9b9bddd5-f192-4b4c-81ee-b523bf9c8705\") " pod="hostpath-provisioner/csi-hostpathplugin-74hx2" Dec 06 03:36:33 crc kubenswrapper[4980]: I1206 03:36:33.863746 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/05aaf766-565d-4b05-97ec-6c8078d33f85-images\") pod \"machine-config-operator-74547568cd-49749\" (UID: \"05aaf766-565d-4b05-97ec-6c8078d33f85\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-49749" Dec 06 03:36:33 crc kubenswrapper[4980]: I1206 03:36:33.863764 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/2692cb89-853a-4c4e-a6a6-570673db6f14-signing-cabundle\") pod \"service-ca-9c57cc56f-qpf6b\" (UID: \"2692cb89-853a-4c4e-a6a6-570673db6f14\") " pod="openshift-service-ca/service-ca-9c57cc56f-qpf6b" Dec 06 03:36:33 crc kubenswrapper[4980]: I1206 03:36:33.863916 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/b25ba819-6e4e-4ca0-a570-c2726b8fbec5-registry-certificates\") pod \"image-registry-697d97f7c8-qw2fb\" (UID: \"b25ba819-6e4e-4ca0-a570-c2726b8fbec5\") " pod="openshift-image-registry/image-registry-697d97f7c8-qw2fb" Dec 06 03:36:33 crc kubenswrapper[4980]: I1206 03:36:33.864042 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c75dl\" (UniqueName: \"kubernetes.io/projected/e1e4e0b4-0bfe-4b01-9c4e-710ba0ec03bd-kube-api-access-c75dl\") pod \"package-server-manager-789f6589d5-qcn9f\" (UID: \"e1e4e0b4-0bfe-4b01-9c4e-710ba0ec03bd\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-qcn9f" Dec 06 03:36:33 crc kubenswrapper[4980]: I1206 03:36:33.864098 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wn452\" (UniqueName: \"kubernetes.io/projected/2692cb89-853a-4c4e-a6a6-570673db6f14-kube-api-access-wn452\") pod \"service-ca-9c57cc56f-qpf6b\" (UID: \"2692cb89-853a-4c4e-a6a6-570673db6f14\") " pod="openshift-service-ca/service-ca-9c57cc56f-qpf6b" Dec 06 03:36:33 crc kubenswrapper[4980]: I1206 03:36:33.864121 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hzvwm\" (UniqueName: \"kubernetes.io/projected/949f499a-5680-4fec-ab59-5ced4f24ba2f-kube-api-access-hzvwm\") pod \"packageserver-d55dfcdfc-gjdsq\" (UID: \"949f499a-5680-4fec-ab59-5ced4f24ba2f\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-gjdsq" Dec 06 03:36:33 crc kubenswrapper[4980]: I1206 03:36:33.864141 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/9476a6a0-4d40-4c85-9bd1-2162c971e3dd-profile-collector-cert\") pod \"olm-operator-6b444d44fb-sq6bp\" (UID: \"9476a6a0-4d40-4c85-9bd1-2162c971e3dd\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-sq6bp" Dec 06 03:36:33 crc kubenswrapper[4980]: I1206 03:36:33.864160 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/8be432d7-9e4b-4452-81df-91998326f92b-config-volume\") pod \"dns-default-4dcjh\" (UID: \"8be432d7-9e4b-4452-81df-91998326f92b\") " pod="openshift-dns/dns-default-4dcjh" Dec 06 03:36:33 crc kubenswrapper[4980]: I1206 03:36:33.864179 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/a7ee597a-a4fc-45ef-841f-1d4f666ccfcf-srv-cert\") pod \"catalog-operator-68c6474976-hjrkf\" (UID: \"a7ee597a-a4fc-45ef-841f-1d4f666ccfcf\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-hjrkf" Dec 06 03:36:33 crc kubenswrapper[4980]: I1206 03:36:33.864200 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/9b9bddd5-f192-4b4c-81ee-b523bf9c8705-mountpoint-dir\") pod \"csi-hostpathplugin-74hx2\" (UID: \"9b9bddd5-f192-4b4c-81ee-b523bf9c8705\") " pod="hostpath-provisioner/csi-hostpathplugin-74hx2" Dec 06 03:36:33 crc kubenswrapper[4980]: I1206 03:36:33.864235 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b25ba819-6e4e-4ca0-a570-c2726b8fbec5-trusted-ca\") pod \"image-registry-697d97f7c8-qw2fb\" (UID: \"b25ba819-6e4e-4ca0-a570-c2726b8fbec5\") " pod="openshift-image-registry/image-registry-697d97f7c8-qw2fb" Dec 06 03:36:33 crc kubenswrapper[4980]: I1206 03:36:33.864262 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c7146eb0-acc5-4bac-b429-e1e00761e666-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-8jc8f\" (UID: \"c7146eb0-acc5-4bac-b429-e1e00761e666\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-8jc8f" Dec 06 03:36:33 crc kubenswrapper[4980]: I1206 03:36:33.864287 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/949f499a-5680-4fec-ab59-5ced4f24ba2f-tmpfs\") pod \"packageserver-d55dfcdfc-gjdsq\" (UID: \"949f499a-5680-4fec-ab59-5ced4f24ba2f\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-gjdsq" Dec 06 03:36:33 crc kubenswrapper[4980]: I1206 03:36:33.864309 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/acb53ac6-3f09-424c-a830-a1e15eec0d62-metrics-certs\") pod \"router-default-5444994796-pzzkd\" (UID: \"acb53ac6-3f09-424c-a830-a1e15eec0d62\") " pod="openshift-ingress/router-default-5444994796-pzzkd" Dec 06 03:36:33 crc kubenswrapper[4980]: I1206 03:36:33.864333 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a3673c8a-c303-42ac-8565-f03e5d457aa1-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-84ft2\" (UID: \"a3673c8a-c303-42ac-8565-f03e5d457aa1\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-84ft2" Dec 06 03:36:33 crc kubenswrapper[4980]: I1206 03:36:33.864360 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/05aaf766-565d-4b05-97ec-6c8078d33f85-proxy-tls\") pod \"machine-config-operator-74547568cd-49749\" (UID: \"05aaf766-565d-4b05-97ec-6c8078d33f85\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-49749" Dec 06 03:36:33 crc kubenswrapper[4980]: I1206 03:36:33.864412 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/a42aad68-42cf-4264-9c94-8678b0eea1b8-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-pmmcw\" (UID: \"a42aad68-42cf-4264-9c94-8678b0eea1b8\") " pod="openshift-authentication/oauth-openshift-558db77b4-pmmcw" Dec 06 03:36:33 crc kubenswrapper[4980]: I1206 03:36:33.864437 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ccd9c9f8-3473-4f3d-884c-e545a0a502af-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-cjvdn\" (UID: \"ccd9c9f8-3473-4f3d-884c-e545a0a502af\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-cjvdn" Dec 06 03:36:33 crc kubenswrapper[4980]: I1206 03:36:33.864479 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qw2fb\" (UID: \"b25ba819-6e4e-4ca0-a570-c2726b8fbec5\") " pod="openshift-image-registry/image-registry-697d97f7c8-qw2fb" Dec 06 03:36:33 crc kubenswrapper[4980]: I1206 03:36:33.864499 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q2hf9\" (UniqueName: \"kubernetes.io/projected/6d695618-ca31-48d0-a178-24d7ee34a54b-kube-api-access-q2hf9\") pod \"console-operator-58897d9998-kftgf\" (UID: \"6d695618-ca31-48d0-a178-24d7ee34a54b\") " pod="openshift-console-operator/console-operator-58897d9998-kftgf" Dec 06 03:36:33 crc kubenswrapper[4980]: I1206 03:36:33.864572 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/768da7ce-a217-4749-be5b-c1ae2dd4e275-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-lrnhm\" (UID: \"768da7ce-a217-4749-be5b-c1ae2dd4e275\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-lrnhm" Dec 06 03:36:33 crc kubenswrapper[4980]: I1206 03:36:33.864596 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a42aad68-42cf-4264-9c94-8678b0eea1b8-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-pmmcw\" (UID: \"a42aad68-42cf-4264-9c94-8678b0eea1b8\") " pod="openshift-authentication/oauth-openshift-558db77b4-pmmcw" Dec 06 03:36:33 crc kubenswrapper[4980]: I1206 03:36:33.864616 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/acb53ac6-3f09-424c-a830-a1e15eec0d62-default-certificate\") pod \"router-default-5444994796-pzzkd\" (UID: \"acb53ac6-3f09-424c-a830-a1e15eec0d62\") " pod="openshift-ingress/router-default-5444994796-pzzkd" Dec 06 03:36:33 crc kubenswrapper[4980]: I1206 03:36:33.864653 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b2da6efd-2b6d-4ca7-b056-32c4b132becf-serving-cert\") pod \"service-ca-operator-777779d784-r47r5\" (UID: \"b2da6efd-2b6d-4ca7-b056-32c4b132becf\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-r47r5" Dec 06 03:36:33 crc kubenswrapper[4980]: I1206 03:36:33.864675 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c7146eb0-acc5-4bac-b429-e1e00761e666-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-8jc8f\" (UID: \"c7146eb0-acc5-4bac-b429-e1e00761e666\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-8jc8f" Dec 06 03:36:33 crc kubenswrapper[4980]: I1206 03:36:33.864721 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ckcls\" (UniqueName: \"kubernetes.io/projected/acb53ac6-3f09-424c-a830-a1e15eec0d62-kube-api-access-ckcls\") pod \"router-default-5444994796-pzzkd\" (UID: \"acb53ac6-3f09-424c-a830-a1e15eec0d62\") " pod="openshift-ingress/router-default-5444994796-pzzkd" Dec 06 03:36:33 crc kubenswrapper[4980]: I1206 03:36:33.864796 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f08eff3e-89fe-4caf-9ec1-b32ee99f3c10-config\") pod \"kube-apiserver-operator-766d6c64bb-xjtnv\" (UID: \"f08eff3e-89fe-4caf-9ec1-b32ee99f3c10\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-xjtnv" Dec 06 03:36:33 crc kubenswrapper[4980]: I1206 03:36:33.864836 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zswsd\" (UniqueName: \"kubernetes.io/projected/92b5e190-2712-40a7-a551-472805f3c84e-kube-api-access-zswsd\") pod \"machine-config-server-bfp7v\" (UID: \"92b5e190-2712-40a7-a551-472805f3c84e\") " pod="openshift-machine-config-operator/machine-config-server-bfp7v" Dec 06 03:36:33 crc kubenswrapper[4980]: I1206 03:36:33.864864 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n2vnn\" (UniqueName: \"kubernetes.io/projected/d1d006b0-92b0-45d5-9c28-fd43f879b310-kube-api-access-n2vnn\") pod \"multus-admission-controller-857f4d67dd-tlpd8\" (UID: \"d1d006b0-92b0-45d5-9c28-fd43f879b310\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-tlpd8" Dec 06 03:36:33 crc kubenswrapper[4980]: I1206 03:36:33.864907 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/abfe3c42-9229-486f-8c87-8b883f6821e0-proxy-tls\") pod \"machine-config-controller-84d6567774-d6dd4\" (UID: \"abfe3c42-9229-486f-8c87-8b883f6821e0\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-d6dd4" Dec 06 03:36:33 crc kubenswrapper[4980]: I1206 03:36:33.864929 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/acb53ac6-3f09-424c-a830-a1e15eec0d62-stats-auth\") pod \"router-default-5444994796-pzzkd\" (UID: \"acb53ac6-3f09-424c-a830-a1e15eec0d62\") " pod="openshift-ingress/router-default-5444994796-pzzkd" Dec 06 03:36:33 crc kubenswrapper[4980]: I1206 03:36:33.864977 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/a42aad68-42cf-4264-9c94-8678b0eea1b8-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-pmmcw\" (UID: \"a42aad68-42cf-4264-9c94-8678b0eea1b8\") " pod="openshift-authentication/oauth-openshift-558db77b4-pmmcw" Dec 06 03:36:33 crc kubenswrapper[4980]: I1206 03:36:33.864995 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/949f499a-5680-4fec-ab59-5ced4f24ba2f-apiservice-cert\") pod \"packageserver-d55dfcdfc-gjdsq\" (UID: \"949f499a-5680-4fec-ab59-5ced4f24ba2f\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-gjdsq" Dec 06 03:36:33 crc kubenswrapper[4980]: I1206 03:36:33.865091 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/768da7ce-a217-4749-be5b-c1ae2dd4e275-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-lrnhm\" (UID: \"768da7ce-a217-4749-be5b-c1ae2dd4e275\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-lrnhm" Dec 06 03:36:33 crc kubenswrapper[4980]: I1206 03:36:33.866439 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/a42aad68-42cf-4264-9c94-8678b0eea1b8-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-pmmcw\" (UID: \"a42aad68-42cf-4264-9c94-8678b0eea1b8\") " pod="openshift-authentication/oauth-openshift-558db77b4-pmmcw" Dec 06 03:36:33 crc kubenswrapper[4980]: I1206 03:36:33.867553 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/a42aad68-42cf-4264-9c94-8678b0eea1b8-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-pmmcw\" (UID: \"a42aad68-42cf-4264-9c94-8678b0eea1b8\") " pod="openshift-authentication/oauth-openshift-558db77b4-pmmcw" Dec 06 03:36:33 crc kubenswrapper[4980]: I1206 03:36:33.868231 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/a42aad68-42cf-4264-9c94-8678b0eea1b8-audit-policies\") pod \"oauth-openshift-558db77b4-pmmcw\" (UID: \"a42aad68-42cf-4264-9c94-8678b0eea1b8\") " pod="openshift-authentication/oauth-openshift-558db77b4-pmmcw" Dec 06 03:36:33 crc kubenswrapper[4980]: I1206 03:36:33.868109 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/b25ba819-6e4e-4ca0-a570-c2726b8fbec5-installation-pull-secrets\") pod \"image-registry-697d97f7c8-qw2fb\" (UID: \"b25ba819-6e4e-4ca0-a570-c2726b8fbec5\") " pod="openshift-image-registry/image-registry-697d97f7c8-qw2fb" Dec 06 03:36:33 crc kubenswrapper[4980]: I1206 03:36:33.870387 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/dae23bd6-4c2b-441e-b44b-b6977bbd974b-metrics-tls\") pod \"dns-operator-744455d44c-w6bmg\" (UID: \"dae23bd6-4c2b-441e-b44b-b6977bbd974b\") " pod="openshift-dns-operator/dns-operator-744455d44c-w6bmg" Dec 06 03:36:33 crc kubenswrapper[4980]: I1206 03:36:33.870838 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/a42aad68-42cf-4264-9c94-8678b0eea1b8-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-pmmcw\" (UID: \"a42aad68-42cf-4264-9c94-8678b0eea1b8\") " pod="openshift-authentication/oauth-openshift-558db77b4-pmmcw" Dec 06 03:36:33 crc kubenswrapper[4980]: I1206 03:36:33.870906 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6d695618-ca31-48d0-a178-24d7ee34a54b-serving-cert\") pod \"console-operator-58897d9998-kftgf\" (UID: \"6d695618-ca31-48d0-a178-24d7ee34a54b\") " pod="openshift-console-operator/console-operator-58897d9998-kftgf" Dec 06 03:36:33 crc kubenswrapper[4980]: I1206 03:36:33.871867 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/a42aad68-42cf-4264-9c94-8678b0eea1b8-audit-dir\") pod \"oauth-openshift-558db77b4-pmmcw\" (UID: \"a42aad68-42cf-4264-9c94-8678b0eea1b8\") " pod="openshift-authentication/oauth-openshift-558db77b4-pmmcw" Dec 06 03:36:33 crc kubenswrapper[4980]: I1206 03:36:33.871873 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a42aad68-42cf-4264-9c94-8678b0eea1b8-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-pmmcw\" (UID: \"a42aad68-42cf-4264-9c94-8678b0eea1b8\") " pod="openshift-authentication/oauth-openshift-558db77b4-pmmcw" Dec 06 03:36:33 crc kubenswrapper[4980]: I1206 03:36:33.872642 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a3673c8a-c303-42ac-8565-f03e5d457aa1-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-84ft2\" (UID: \"a3673c8a-c303-42ac-8565-f03e5d457aa1\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-84ft2" Dec 06 03:36:33 crc kubenswrapper[4980]: I1206 03:36:33.873475 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/a42aad68-42cf-4264-9c94-8678b0eea1b8-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-pmmcw\" (UID: \"a42aad68-42cf-4264-9c94-8678b0eea1b8\") " pod="openshift-authentication/oauth-openshift-558db77b4-pmmcw" Dec 06 03:36:33 crc kubenswrapper[4980]: I1206 03:36:33.873749 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/a42aad68-42cf-4264-9c94-8678b0eea1b8-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-pmmcw\" (UID: \"a42aad68-42cf-4264-9c94-8678b0eea1b8\") " pod="openshift-authentication/oauth-openshift-558db77b4-pmmcw" Dec 06 03:36:33 crc kubenswrapper[4980]: I1206 03:36:33.875056 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a3673c8a-c303-42ac-8565-f03e5d457aa1-serving-cert\") pod \"authentication-operator-69f744f599-84ft2\" (UID: \"a3673c8a-c303-42ac-8565-f03e5d457aa1\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-84ft2" Dec 06 03:36:33 crc kubenswrapper[4980]: I1206 03:36:33.875456 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b25ba819-6e4e-4ca0-a570-c2726b8fbec5-trusted-ca\") pod \"image-registry-697d97f7c8-qw2fb\" (UID: \"b25ba819-6e4e-4ca0-a570-c2726b8fbec5\") " pod="openshift-image-registry/image-registry-697d97f7c8-qw2fb" Dec 06 03:36:33 crc kubenswrapper[4980]: I1206 03:36:33.875570 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/a42aad68-42cf-4264-9c94-8678b0eea1b8-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-pmmcw\" (UID: \"a42aad68-42cf-4264-9c94-8678b0eea1b8\") " pod="openshift-authentication/oauth-openshift-558db77b4-pmmcw" Dec 06 03:36:33 crc kubenswrapper[4980]: E1206 03:36:33.875950 4980 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 03:36:34.375928513 +0000 UTC m=+153.620679784 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qw2fb" (UID: "b25ba819-6e4e-4ca0-a570-c2726b8fbec5") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 03:36:33 crc kubenswrapper[4980]: I1206 03:36:33.875969 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/b25ba819-6e4e-4ca0-a570-c2726b8fbec5-ca-trust-extracted\") pod \"image-registry-697d97f7c8-qw2fb\" (UID: \"b25ba819-6e4e-4ca0-a570-c2726b8fbec5\") " pod="openshift-image-registry/image-registry-697d97f7c8-qw2fb" Dec 06 03:36:33 crc kubenswrapper[4980]: I1206 03:36:33.876756 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6d695618-ca31-48d0-a178-24d7ee34a54b-config\") pod \"console-operator-58897d9998-kftgf\" (UID: \"6d695618-ca31-48d0-a178-24d7ee34a54b\") " pod="openshift-console-operator/console-operator-58897d9998-kftgf" Dec 06 03:36:33 crc kubenswrapper[4980]: I1206 03:36:33.877983 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/b25ba819-6e4e-4ca0-a570-c2726b8fbec5-registry-tls\") pod \"image-registry-697d97f7c8-qw2fb\" (UID: \"b25ba819-6e4e-4ca0-a570-c2726b8fbec5\") " pod="openshift-image-registry/image-registry-697d97f7c8-qw2fb" Dec 06 03:36:33 crc kubenswrapper[4980]: I1206 03:36:33.878281 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/6d695618-ca31-48d0-a178-24d7ee34a54b-trusted-ca\") pod \"console-operator-58897d9998-kftgf\" (UID: \"6d695618-ca31-48d0-a178-24d7ee34a54b\") " pod="openshift-console-operator/console-operator-58897d9998-kftgf" Dec 06 03:36:33 crc kubenswrapper[4980]: I1206 03:36:33.879118 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a3673c8a-c303-42ac-8565-f03e5d457aa1-service-ca-bundle\") pod \"authentication-operator-69f744f599-84ft2\" (UID: \"a3673c8a-c303-42ac-8565-f03e5d457aa1\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-84ft2" Dec 06 03:36:33 crc kubenswrapper[4980]: I1206 03:36:33.879190 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/a42aad68-42cf-4264-9c94-8678b0eea1b8-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-pmmcw\" (UID: \"a42aad68-42cf-4264-9c94-8678b0eea1b8\") " pod="openshift-authentication/oauth-openshift-558db77b4-pmmcw" Dec 06 03:36:33 crc kubenswrapper[4980]: I1206 03:36:33.885137 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a3673c8a-c303-42ac-8565-f03e5d457aa1-config\") pod \"authentication-operator-69f744f599-84ft2\" (UID: \"a3673c8a-c303-42ac-8565-f03e5d457aa1\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-84ft2" Dec 06 03:36:33 crc kubenswrapper[4980]: I1206 03:36:33.884759 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/a42aad68-42cf-4264-9c94-8678b0eea1b8-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-pmmcw\" (UID: \"a42aad68-42cf-4264-9c94-8678b0eea1b8\") " pod="openshift-authentication/oauth-openshift-558db77b4-pmmcw" Dec 06 03:36:33 crc kubenswrapper[4980]: I1206 03:36:33.886670 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/a42aad68-42cf-4264-9c94-8678b0eea1b8-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-pmmcw\" (UID: \"a42aad68-42cf-4264-9c94-8678b0eea1b8\") " pod="openshift-authentication/oauth-openshift-558db77b4-pmmcw" Dec 06 03:36:33 crc kubenswrapper[4980]: I1206 03:36:33.888371 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/768da7ce-a217-4749-be5b-c1ae2dd4e275-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-lrnhm\" (UID: \"768da7ce-a217-4749-be5b-c1ae2dd4e275\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-lrnhm" Dec 06 03:36:33 crc kubenswrapper[4980]: I1206 03:36:33.889985 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/a42aad68-42cf-4264-9c94-8678b0eea1b8-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-pmmcw\" (UID: \"a42aad68-42cf-4264-9c94-8678b0eea1b8\") " pod="openshift-authentication/oauth-openshift-558db77b4-pmmcw" Dec 06 03:36:33 crc kubenswrapper[4980]: I1206 03:36:33.965352 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 03:36:33 crc kubenswrapper[4980]: I1206 03:36:33.965477 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/949f499a-5680-4fec-ab59-5ced4f24ba2f-tmpfs\") pod \"packageserver-d55dfcdfc-gjdsq\" (UID: \"949f499a-5680-4fec-ab59-5ced4f24ba2f\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-gjdsq" Dec 06 03:36:33 crc kubenswrapper[4980]: E1206 03:36:33.965505 4980 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 03:36:34.465487308 +0000 UTC m=+153.710238579 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 03:36:33 crc kubenswrapper[4980]: I1206 03:36:33.965544 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/acb53ac6-3f09-424c-a830-a1e15eec0d62-metrics-certs\") pod \"router-default-5444994796-pzzkd\" (UID: \"acb53ac6-3f09-424c-a830-a1e15eec0d62\") " pod="openshift-ingress/router-default-5444994796-pzzkd" Dec 06 03:36:33 crc kubenswrapper[4980]: I1206 03:36:33.965568 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c7146eb0-acc5-4bac-b429-e1e00761e666-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-8jc8f\" (UID: \"c7146eb0-acc5-4bac-b429-e1e00761e666\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-8jc8f" Dec 06 03:36:33 crc kubenswrapper[4980]: I1206 03:36:33.965587 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/05aaf766-565d-4b05-97ec-6c8078d33f85-proxy-tls\") pod \"machine-config-operator-74547568cd-49749\" (UID: \"05aaf766-565d-4b05-97ec-6c8078d33f85\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-49749" Dec 06 03:36:33 crc kubenswrapper[4980]: I1206 03:36:33.965607 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ccd9c9f8-3473-4f3d-884c-e545a0a502af-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-cjvdn\" (UID: \"ccd9c9f8-3473-4f3d-884c-e545a0a502af\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-cjvdn" Dec 06 03:36:33 crc kubenswrapper[4980]: I1206 03:36:33.965627 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qw2fb\" (UID: \"b25ba819-6e4e-4ca0-a570-c2726b8fbec5\") " pod="openshift-image-registry/image-registry-697d97f7c8-qw2fb" Dec 06 03:36:33 crc kubenswrapper[4980]: I1206 03:36:33.965656 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/acb53ac6-3f09-424c-a830-a1e15eec0d62-default-certificate\") pod \"router-default-5444994796-pzzkd\" (UID: \"acb53ac6-3f09-424c-a830-a1e15eec0d62\") " pod="openshift-ingress/router-default-5444994796-pzzkd" Dec 06 03:36:33 crc kubenswrapper[4980]: I1206 03:36:33.965673 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b2da6efd-2b6d-4ca7-b056-32c4b132becf-serving-cert\") pod \"service-ca-operator-777779d784-r47r5\" (UID: \"b2da6efd-2b6d-4ca7-b056-32c4b132becf\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-r47r5" Dec 06 03:36:33 crc kubenswrapper[4980]: I1206 03:36:33.965688 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c7146eb0-acc5-4bac-b429-e1e00761e666-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-8jc8f\" (UID: \"c7146eb0-acc5-4bac-b429-e1e00761e666\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-8jc8f" Dec 06 03:36:33 crc kubenswrapper[4980]: I1206 03:36:33.965707 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ckcls\" (UniqueName: \"kubernetes.io/projected/acb53ac6-3f09-424c-a830-a1e15eec0d62-kube-api-access-ckcls\") pod \"router-default-5444994796-pzzkd\" (UID: \"acb53ac6-3f09-424c-a830-a1e15eec0d62\") " pod="openshift-ingress/router-default-5444994796-pzzkd" Dec 06 03:36:33 crc kubenswrapper[4980]: I1206 03:36:33.965725 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f08eff3e-89fe-4caf-9ec1-b32ee99f3c10-config\") pod \"kube-apiserver-operator-766d6c64bb-xjtnv\" (UID: \"f08eff3e-89fe-4caf-9ec1-b32ee99f3c10\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-xjtnv" Dec 06 03:36:33 crc kubenswrapper[4980]: I1206 03:36:33.965749 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zswsd\" (UniqueName: \"kubernetes.io/projected/92b5e190-2712-40a7-a551-472805f3c84e-kube-api-access-zswsd\") pod \"machine-config-server-bfp7v\" (UID: \"92b5e190-2712-40a7-a551-472805f3c84e\") " pod="openshift-machine-config-operator/machine-config-server-bfp7v" Dec 06 03:36:33 crc kubenswrapper[4980]: I1206 03:36:33.965785 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n2vnn\" (UniqueName: \"kubernetes.io/projected/d1d006b0-92b0-45d5-9c28-fd43f879b310-kube-api-access-n2vnn\") pod \"multus-admission-controller-857f4d67dd-tlpd8\" (UID: \"d1d006b0-92b0-45d5-9c28-fd43f879b310\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-tlpd8" Dec 06 03:36:33 crc kubenswrapper[4980]: I1206 03:36:33.965806 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/abfe3c42-9229-486f-8c87-8b883f6821e0-proxy-tls\") pod \"machine-config-controller-84d6567774-d6dd4\" (UID: \"abfe3c42-9229-486f-8c87-8b883f6821e0\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-d6dd4" Dec 06 03:36:33 crc kubenswrapper[4980]: I1206 03:36:33.965824 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/acb53ac6-3f09-424c-a830-a1e15eec0d62-stats-auth\") pod \"router-default-5444994796-pzzkd\" (UID: \"acb53ac6-3f09-424c-a830-a1e15eec0d62\") " pod="openshift-ingress/router-default-5444994796-pzzkd" Dec 06 03:36:33 crc kubenswrapper[4980]: I1206 03:36:33.965841 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/949f499a-5680-4fec-ab59-5ced4f24ba2f-apiservice-cert\") pod \"packageserver-d55dfcdfc-gjdsq\" (UID: \"949f499a-5680-4fec-ab59-5ced4f24ba2f\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-gjdsq" Dec 06 03:36:33 crc kubenswrapper[4980]: I1206 03:36:33.965893 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/f08eff3e-89fe-4caf-9ec1-b32ee99f3c10-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-xjtnv\" (UID: \"f08eff3e-89fe-4caf-9ec1-b32ee99f3c10\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-xjtnv" Dec 06 03:36:33 crc kubenswrapper[4980]: I1206 03:36:33.965915 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dp275\" (UniqueName: \"kubernetes.io/projected/ccd9c9f8-3473-4f3d-884c-e545a0a502af-kube-api-access-dp275\") pod \"kube-storage-version-migrator-operator-b67b599dd-cjvdn\" (UID: \"ccd9c9f8-3473-4f3d-884c-e545a0a502af\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-cjvdn" Dec 06 03:36:33 crc kubenswrapper[4980]: I1206 03:36:33.965930 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mfkfg\" (UniqueName: \"kubernetes.io/projected/d7fb8e4e-88e7-4cc9-baf9-6e217813b28d-kube-api-access-mfkfg\") pod \"migrator-59844c95c7-56s8p\" (UID: \"d7fb8e4e-88e7-4cc9-baf9-6e217813b28d\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-56s8p" Dec 06 03:36:33 crc kubenswrapper[4980]: I1206 03:36:33.965954 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/acb53ac6-3f09-424c-a830-a1e15eec0d62-service-ca-bundle\") pod \"router-default-5444994796-pzzkd\" (UID: \"acb53ac6-3f09-424c-a830-a1e15eec0d62\") " pod="openshift-ingress/router-default-5444994796-pzzkd" Dec 06 03:36:33 crc kubenswrapper[4980]: I1206 03:36:33.965962 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/949f499a-5680-4fec-ab59-5ced4f24ba2f-tmpfs\") pod \"packageserver-d55dfcdfc-gjdsq\" (UID: \"949f499a-5680-4fec-ab59-5ced4f24ba2f\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-gjdsq" Dec 06 03:36:33 crc kubenswrapper[4980]: I1206 03:36:33.965976 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t52g4\" (UniqueName: \"kubernetes.io/projected/3ee9a5b6-d84f-46d7-b3c2-aab85c634bd4-kube-api-access-t52g4\") pod \"ingress-canary-799j5\" (UID: \"3ee9a5b6-d84f-46d7-b3c2-aab85c634bd4\") " pod="openshift-ingress-canary/ingress-canary-799j5" Dec 06 03:36:33 crc kubenswrapper[4980]: I1206 03:36:33.965996 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/46290605-e114-4beb-b1c9-b0389f20f5f3-bound-sa-token\") pod \"ingress-operator-5b745b69d9-27gc7\" (UID: \"46290605-e114-4beb-b1c9-b0389f20f5f3\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-27gc7" Dec 06 03:36:33 crc kubenswrapper[4980]: I1206 03:36:33.966017 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/92b5e190-2712-40a7-a551-472805f3c84e-certs\") pod \"machine-config-server-bfp7v\" (UID: \"92b5e190-2712-40a7-a551-472805f3c84e\") " pod="openshift-machine-config-operator/machine-config-server-bfp7v" Dec 06 03:36:33 crc kubenswrapper[4980]: I1206 03:36:33.966038 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/92b5e190-2712-40a7-a551-472805f3c84e-node-bootstrap-token\") pod \"machine-config-server-bfp7v\" (UID: \"92b5e190-2712-40a7-a551-472805f3c84e\") " pod="openshift-machine-config-operator/machine-config-server-bfp7v" Dec 06 03:36:33 crc kubenswrapper[4980]: I1206 03:36:33.966060 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/0a45bc54-be9c-4c4f-9cae-dc369d25be33-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-spd2v\" (UID: \"0a45bc54-be9c-4c4f-9cae-dc369d25be33\") " pod="openshift-marketplace/marketplace-operator-79b997595-spd2v" Dec 06 03:36:33 crc kubenswrapper[4980]: I1206 03:36:33.966082 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wr5kj\" (UniqueName: \"kubernetes.io/projected/05aaf766-565d-4b05-97ec-6c8078d33f85-kube-api-access-wr5kj\") pod \"machine-config-operator-74547568cd-49749\" (UID: \"05aaf766-565d-4b05-97ec-6c8078d33f85\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-49749" Dec 06 03:36:33 crc kubenswrapper[4980]: I1206 03:36:33.966103 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/8be432d7-9e4b-4452-81df-91998326f92b-metrics-tls\") pod \"dns-default-4dcjh\" (UID: \"8be432d7-9e4b-4452-81df-91998326f92b\") " pod="openshift-dns/dns-default-4dcjh" Dec 06 03:36:33 crc kubenswrapper[4980]: I1206 03:36:33.966122 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/0a45bc54-be9c-4c4f-9cae-dc369d25be33-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-spd2v\" (UID: \"0a45bc54-be9c-4c4f-9cae-dc369d25be33\") " pod="openshift-marketplace/marketplace-operator-79b997595-spd2v" Dec 06 03:36:33 crc kubenswrapper[4980]: I1206 03:36:33.966142 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/05aaf766-565d-4b05-97ec-6c8078d33f85-auth-proxy-config\") pod \"machine-config-operator-74547568cd-49749\" (UID: \"05aaf766-565d-4b05-97ec-6c8078d33f85\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-49749" Dec 06 03:36:33 crc kubenswrapper[4980]: I1206 03:36:33.966163 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ccd9c9f8-3473-4f3d-884c-e545a0a502af-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-cjvdn\" (UID: \"ccd9c9f8-3473-4f3d-884c-e545a0a502af\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-cjvdn" Dec 06 03:36:33 crc kubenswrapper[4980]: I1206 03:36:33.966181 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b2da6efd-2b6d-4ca7-b056-32c4b132becf-config\") pod \"service-ca-operator-777779d784-r47r5\" (UID: \"b2da6efd-2b6d-4ca7-b056-32c4b132becf\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-r47r5" Dec 06 03:36:33 crc kubenswrapper[4980]: I1206 03:36:33.966238 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f08eff3e-89fe-4caf-9ec1-b32ee99f3c10-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-xjtnv\" (UID: \"f08eff3e-89fe-4caf-9ec1-b32ee99f3c10\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-xjtnv" Dec 06 03:36:33 crc kubenswrapper[4980]: I1206 03:36:33.966265 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vcwpf\" (UniqueName: \"kubernetes.io/projected/9476a6a0-4d40-4c85-9bd1-2162c971e3dd-kube-api-access-vcwpf\") pod \"olm-operator-6b444d44fb-sq6bp\" (UID: \"9476a6a0-4d40-4c85-9bd1-2162c971e3dd\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-sq6bp" Dec 06 03:36:33 crc kubenswrapper[4980]: I1206 03:36:33.966288 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t26fr\" (UniqueName: \"kubernetes.io/projected/b2da6efd-2b6d-4ca7-b056-32c4b132becf-kube-api-access-t26fr\") pod \"service-ca-operator-777779d784-r47r5\" (UID: \"b2da6efd-2b6d-4ca7-b056-32c4b132becf\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-r47r5" Dec 06 03:36:33 crc kubenswrapper[4980]: I1206 03:36:33.966310 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/9476a6a0-4d40-4c85-9bd1-2162c971e3dd-srv-cert\") pod \"olm-operator-6b444d44fb-sq6bp\" (UID: \"9476a6a0-4d40-4c85-9bd1-2162c971e3dd\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-sq6bp" Dec 06 03:36:33 crc kubenswrapper[4980]: I1206 03:36:33.966364 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/46290605-e114-4beb-b1c9-b0389f20f5f3-metrics-tls\") pod \"ingress-operator-5b745b69d9-27gc7\" (UID: \"46290605-e114-4beb-b1c9-b0389f20f5f3\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-27gc7" Dec 06 03:36:33 crc kubenswrapper[4980]: I1206 03:36:33.966385 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/46290605-e114-4beb-b1c9-b0389f20f5f3-trusted-ca\") pod \"ingress-operator-5b745b69d9-27gc7\" (UID: \"46290605-e114-4beb-b1c9-b0389f20f5f3\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-27gc7" Dec 06 03:36:33 crc kubenswrapper[4980]: I1206 03:36:33.966417 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z4m5n\" (UniqueName: \"kubernetes.io/projected/abfe3c42-9229-486f-8c87-8b883f6821e0-kube-api-access-z4m5n\") pod \"machine-config-controller-84d6567774-d6dd4\" (UID: \"abfe3c42-9229-486f-8c87-8b883f6821e0\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-d6dd4" Dec 06 03:36:33 crc kubenswrapper[4980]: I1206 03:36:33.966442 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/e81d9a6d-991f-46b0-aba5-6f76a5b1031a-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-cqk6j\" (UID: \"e81d9a6d-991f-46b0-aba5-6f76a5b1031a\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-cqk6j" Dec 06 03:36:33 crc kubenswrapper[4980]: I1206 03:36:33.966472 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/03704eb9-227d-4985-87cf-acd66ae76037-config-volume\") pod \"collect-profiles-29416530-9hm9m\" (UID: \"03704eb9-227d-4985-87cf-acd66ae76037\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416530-9hm9m" Dec 06 03:36:33 crc kubenswrapper[4980]: I1206 03:36:33.966493 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/e1e4e0b4-0bfe-4b01-9c4e-710ba0ec03bd-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-qcn9f\" (UID: \"e1e4e0b4-0bfe-4b01-9c4e-710ba0ec03bd\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-qcn9f" Dec 06 03:36:33 crc kubenswrapper[4980]: I1206 03:36:33.966532 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/abfe3c42-9229-486f-8c87-8b883f6821e0-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-d6dd4\" (UID: \"abfe3c42-9229-486f-8c87-8b883f6821e0\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-d6dd4" Dec 06 03:36:33 crc kubenswrapper[4980]: I1206 03:36:33.966553 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f08eff3e-89fe-4caf-9ec1-b32ee99f3c10-config\") pod \"kube-apiserver-operator-766d6c64bb-xjtnv\" (UID: \"f08eff3e-89fe-4caf-9ec1-b32ee99f3c10\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-xjtnv" Dec 06 03:36:33 crc kubenswrapper[4980]: E1206 03:36:33.966807 4980 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 03:36:34.466794376 +0000 UTC m=+153.711545647 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qw2fb" (UID: "b25ba819-6e4e-4ca0-a570-c2726b8fbec5") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 03:36:33 crc kubenswrapper[4980]: I1206 03:36:33.967829 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c7146eb0-acc5-4bac-b429-e1e00761e666-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-8jc8f\" (UID: \"c7146eb0-acc5-4bac-b429-e1e00761e666\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-8jc8f" Dec 06 03:36:33 crc kubenswrapper[4980]: I1206 03:36:33.966554 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f2555ff9-efdb-449a-9f88-67a3c80ad0cf-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-jp25g\" (UID: \"f2555ff9-efdb-449a-9f88-67a3c80ad0cf\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-jp25g" Dec 06 03:36:33 crc kubenswrapper[4980]: I1206 03:36:33.968143 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9p5zn\" (UniqueName: \"kubernetes.io/projected/46290605-e114-4beb-b1c9-b0389f20f5f3-kube-api-access-9p5zn\") pod \"ingress-operator-5b745b69d9-27gc7\" (UID: \"46290605-e114-4beb-b1c9-b0389f20f5f3\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-27gc7" Dec 06 03:36:33 crc kubenswrapper[4980]: I1206 03:36:33.968223 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x6x86\" (UniqueName: \"kubernetes.io/projected/a7ee597a-a4fc-45ef-841f-1d4f666ccfcf-kube-api-access-x6x86\") pod \"catalog-operator-68c6474976-hjrkf\" (UID: \"a7ee597a-a4fc-45ef-841f-1d4f666ccfcf\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-hjrkf" Dec 06 03:36:33 crc kubenswrapper[4980]: I1206 03:36:33.968289 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/3ee9a5b6-d84f-46d7-b3c2-aab85c634bd4-cert\") pod \"ingress-canary-799j5\" (UID: \"3ee9a5b6-d84f-46d7-b3c2-aab85c634bd4\") " pod="openshift-ingress-canary/ingress-canary-799j5" Dec 06 03:36:33 crc kubenswrapper[4980]: I1206 03:36:33.968326 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/9b9bddd5-f192-4b4c-81ee-b523bf9c8705-plugins-dir\") pod \"csi-hostpathplugin-74hx2\" (UID: \"9b9bddd5-f192-4b4c-81ee-b523bf9c8705\") " pod="hostpath-provisioner/csi-hostpathplugin-74hx2" Dec 06 03:36:33 crc kubenswrapper[4980]: I1206 03:36:33.968366 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/c7146eb0-acc5-4bac-b429-e1e00761e666-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-8jc8f\" (UID: \"c7146eb0-acc5-4bac-b429-e1e00761e666\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-8jc8f" Dec 06 03:36:33 crc kubenswrapper[4980]: I1206 03:36:33.968421 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/2692cb89-853a-4c4e-a6a6-570673db6f14-signing-key\") pod \"service-ca-9c57cc56f-qpf6b\" (UID: \"2692cb89-853a-4c4e-a6a6-570673db6f14\") " pod="openshift-service-ca/service-ca-9c57cc56f-qpf6b" Dec 06 03:36:33 crc kubenswrapper[4980]: I1206 03:36:33.968444 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/9b9bddd5-f192-4b4c-81ee-b523bf9c8705-socket-dir\") pod \"csi-hostpathplugin-74hx2\" (UID: \"9b9bddd5-f192-4b4c-81ee-b523bf9c8705\") " pod="hostpath-provisioner/csi-hostpathplugin-74hx2" Dec 06 03:36:33 crc kubenswrapper[4980]: I1206 03:36:33.968475 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/9b9bddd5-f192-4b4c-81ee-b523bf9c8705-registration-dir\") pod \"csi-hostpathplugin-74hx2\" (UID: \"9b9bddd5-f192-4b4c-81ee-b523bf9c8705\") " pod="hostpath-provisioner/csi-hostpathplugin-74hx2" Dec 06 03:36:33 crc kubenswrapper[4980]: I1206 03:36:33.968758 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/949f499a-5680-4fec-ab59-5ced4f24ba2f-webhook-cert\") pod \"packageserver-d55dfcdfc-gjdsq\" (UID: \"949f499a-5680-4fec-ab59-5ced4f24ba2f\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-gjdsq" Dec 06 03:36:33 crc kubenswrapper[4980]: I1206 03:36:33.968786 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f2555ff9-efdb-449a-9f88-67a3c80ad0cf-config\") pod \"kube-controller-manager-operator-78b949d7b-jp25g\" (UID: \"f2555ff9-efdb-449a-9f88-67a3c80ad0cf\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-jp25g" Dec 06 03:36:33 crc kubenswrapper[4980]: I1206 03:36:33.968808 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vg8kf\" (UniqueName: \"kubernetes.io/projected/e81d9a6d-991f-46b0-aba5-6f76a5b1031a-kube-api-access-vg8kf\") pod \"control-plane-machine-set-operator-78cbb6b69f-cqk6j\" (UID: \"e81d9a6d-991f-46b0-aba5-6f76a5b1031a\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-cqk6j" Dec 06 03:36:33 crc kubenswrapper[4980]: I1206 03:36:33.968832 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/03704eb9-227d-4985-87cf-acd66ae76037-secret-volume\") pod \"collect-profiles-29416530-9hm9m\" (UID: \"03704eb9-227d-4985-87cf-acd66ae76037\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416530-9hm9m" Dec 06 03:36:33 crc kubenswrapper[4980]: I1206 03:36:33.968853 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f5l5z\" (UniqueName: \"kubernetes.io/projected/0a45bc54-be9c-4c4f-9cae-dc369d25be33-kube-api-access-f5l5z\") pod \"marketplace-operator-79b997595-spd2v\" (UID: \"0a45bc54-be9c-4c4f-9cae-dc369d25be33\") " pod="openshift-marketplace/marketplace-operator-79b997595-spd2v" Dec 06 03:36:33 crc kubenswrapper[4980]: I1206 03:36:33.968874 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/9b9bddd5-f192-4b4c-81ee-b523bf9c8705-csi-data-dir\") pod \"csi-hostpathplugin-74hx2\" (UID: \"9b9bddd5-f192-4b4c-81ee-b523bf9c8705\") " pod="hostpath-provisioner/csi-hostpathplugin-74hx2" Dec 06 03:36:33 crc kubenswrapper[4980]: I1206 03:36:33.968896 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/a7ee597a-a4fc-45ef-841f-1d4f666ccfcf-profile-collector-cert\") pod \"catalog-operator-68c6474976-hjrkf\" (UID: \"a7ee597a-a4fc-45ef-841f-1d4f666ccfcf\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-hjrkf" Dec 06 03:36:33 crc kubenswrapper[4980]: I1206 03:36:33.968912 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/0a45bc54-be9c-4c4f-9cae-dc369d25be33-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-spd2v\" (UID: \"0a45bc54-be9c-4c4f-9cae-dc369d25be33\") " pod="openshift-marketplace/marketplace-operator-79b997595-spd2v" Dec 06 03:36:33 crc kubenswrapper[4980]: I1206 03:36:33.968926 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/f2555ff9-efdb-449a-9f88-67a3c80ad0cf-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-jp25g\" (UID: \"f2555ff9-efdb-449a-9f88-67a3c80ad0cf\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-jp25g" Dec 06 03:36:33 crc kubenswrapper[4980]: I1206 03:36:33.969017 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b6q5n\" (UniqueName: \"kubernetes.io/projected/8be432d7-9e4b-4452-81df-91998326f92b-kube-api-access-b6q5n\") pod \"dns-default-4dcjh\" (UID: \"8be432d7-9e4b-4452-81df-91998326f92b\") " pod="openshift-dns/dns-default-4dcjh" Dec 06 03:36:33 crc kubenswrapper[4980]: I1206 03:36:33.969045 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/d1d006b0-92b0-45d5-9c28-fd43f879b310-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-tlpd8\" (UID: \"d1d006b0-92b0-45d5-9c28-fd43f879b310\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-tlpd8" Dec 06 03:36:33 crc kubenswrapper[4980]: I1206 03:36:33.969072 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2w6nc\" (UniqueName: \"kubernetes.io/projected/03704eb9-227d-4985-87cf-acd66ae76037-kube-api-access-2w6nc\") pod \"collect-profiles-29416530-9hm9m\" (UID: \"03704eb9-227d-4985-87cf-acd66ae76037\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416530-9hm9m" Dec 06 03:36:33 crc kubenswrapper[4980]: I1206 03:36:33.969093 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pl4qc\" (UniqueName: \"kubernetes.io/projected/9b9bddd5-f192-4b4c-81ee-b523bf9c8705-kube-api-access-pl4qc\") pod \"csi-hostpathplugin-74hx2\" (UID: \"9b9bddd5-f192-4b4c-81ee-b523bf9c8705\") " pod="hostpath-provisioner/csi-hostpathplugin-74hx2" Dec 06 03:36:33 crc kubenswrapper[4980]: I1206 03:36:33.969117 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/05aaf766-565d-4b05-97ec-6c8078d33f85-images\") pod \"machine-config-operator-74547568cd-49749\" (UID: \"05aaf766-565d-4b05-97ec-6c8078d33f85\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-49749" Dec 06 03:36:33 crc kubenswrapper[4980]: I1206 03:36:33.969148 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/2692cb89-853a-4c4e-a6a6-570673db6f14-signing-cabundle\") pod \"service-ca-9c57cc56f-qpf6b\" (UID: \"2692cb89-853a-4c4e-a6a6-570673db6f14\") " pod="openshift-service-ca/service-ca-9c57cc56f-qpf6b" Dec 06 03:36:33 crc kubenswrapper[4980]: I1206 03:36:33.969205 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c75dl\" (UniqueName: \"kubernetes.io/projected/e1e4e0b4-0bfe-4b01-9c4e-710ba0ec03bd-kube-api-access-c75dl\") pod \"package-server-manager-789f6589d5-qcn9f\" (UID: \"e1e4e0b4-0bfe-4b01-9c4e-710ba0ec03bd\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-qcn9f" Dec 06 03:36:33 crc kubenswrapper[4980]: I1206 03:36:33.969238 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wn452\" (UniqueName: \"kubernetes.io/projected/2692cb89-853a-4c4e-a6a6-570673db6f14-kube-api-access-wn452\") pod \"service-ca-9c57cc56f-qpf6b\" (UID: \"2692cb89-853a-4c4e-a6a6-570673db6f14\") " pod="openshift-service-ca/service-ca-9c57cc56f-qpf6b" Dec 06 03:36:33 crc kubenswrapper[4980]: I1206 03:36:33.969259 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hzvwm\" (UniqueName: \"kubernetes.io/projected/949f499a-5680-4fec-ab59-5ced4f24ba2f-kube-api-access-hzvwm\") pod \"packageserver-d55dfcdfc-gjdsq\" (UID: \"949f499a-5680-4fec-ab59-5ced4f24ba2f\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-gjdsq" Dec 06 03:36:33 crc kubenswrapper[4980]: I1206 03:36:33.969280 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/9476a6a0-4d40-4c85-9bd1-2162c971e3dd-profile-collector-cert\") pod \"olm-operator-6b444d44fb-sq6bp\" (UID: \"9476a6a0-4d40-4c85-9bd1-2162c971e3dd\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-sq6bp" Dec 06 03:36:33 crc kubenswrapper[4980]: I1206 03:36:33.969305 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/8be432d7-9e4b-4452-81df-91998326f92b-config-volume\") pod \"dns-default-4dcjh\" (UID: \"8be432d7-9e4b-4452-81df-91998326f92b\") " pod="openshift-dns/dns-default-4dcjh" Dec 06 03:36:33 crc kubenswrapper[4980]: I1206 03:36:33.969324 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/a7ee597a-a4fc-45ef-841f-1d4f666ccfcf-srv-cert\") pod \"catalog-operator-68c6474976-hjrkf\" (UID: \"a7ee597a-a4fc-45ef-841f-1d4f666ccfcf\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-hjrkf" Dec 06 03:36:33 crc kubenswrapper[4980]: I1206 03:36:33.969342 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/9b9bddd5-f192-4b4c-81ee-b523bf9c8705-mountpoint-dir\") pod \"csi-hostpathplugin-74hx2\" (UID: \"9b9bddd5-f192-4b4c-81ee-b523bf9c8705\") " pod="hostpath-provisioner/csi-hostpathplugin-74hx2" Dec 06 03:36:33 crc kubenswrapper[4980]: I1206 03:36:33.969615 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/9b9bddd5-f192-4b4c-81ee-b523bf9c8705-mountpoint-dir\") pod \"csi-hostpathplugin-74hx2\" (UID: \"9b9bddd5-f192-4b4c-81ee-b523bf9c8705\") " pod="hostpath-provisioner/csi-hostpathplugin-74hx2" Dec 06 03:36:33 crc kubenswrapper[4980]: I1206 03:36:33.973253 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/acb53ac6-3f09-424c-a830-a1e15eec0d62-metrics-certs\") pod \"router-default-5444994796-pzzkd\" (UID: \"acb53ac6-3f09-424c-a830-a1e15eec0d62\") " pod="openshift-ingress/router-default-5444994796-pzzkd" Dec 06 03:36:33 crc kubenswrapper[4980]: I1206 03:36:33.978697 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f2555ff9-efdb-449a-9f88-67a3c80ad0cf-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-jp25g\" (UID: \"f2555ff9-efdb-449a-9f88-67a3c80ad0cf\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-jp25g" Dec 06 03:36:33 crc kubenswrapper[4980]: I1206 03:36:33.980022 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b2da6efd-2b6d-4ca7-b056-32c4b132becf-serving-cert\") pod \"service-ca-operator-777779d784-r47r5\" (UID: \"b2da6efd-2b6d-4ca7-b056-32c4b132becf\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-r47r5" Dec 06 03:36:33 crc kubenswrapper[4980]: I1206 03:36:33.980072 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q2hf9\" (UniqueName: \"kubernetes.io/projected/6d695618-ca31-48d0-a178-24d7ee34a54b-kube-api-access-q2hf9\") pod \"console-operator-58897d9998-kftgf\" (UID: \"6d695618-ca31-48d0-a178-24d7ee34a54b\") " pod="openshift-console-operator/console-operator-58897d9998-kftgf" Dec 06 03:36:33 crc kubenswrapper[4980]: I1206 03:36:33.980911 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ccd9c9f8-3473-4f3d-884c-e545a0a502af-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-cjvdn\" (UID: \"ccd9c9f8-3473-4f3d-884c-e545a0a502af\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-cjvdn" Dec 06 03:36:33 crc kubenswrapper[4980]: I1206 03:36:33.981285 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/03704eb9-227d-4985-87cf-acd66ae76037-config-volume\") pod \"collect-profiles-29416530-9hm9m\" (UID: \"03704eb9-227d-4985-87cf-acd66ae76037\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416530-9hm9m" Dec 06 03:36:33 crc kubenswrapper[4980]: I1206 03:36:33.982010 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/abfe3c42-9229-486f-8c87-8b883f6821e0-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-d6dd4\" (UID: \"abfe3c42-9229-486f-8c87-8b883f6821e0\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-d6dd4" Dec 06 03:36:33 crc kubenswrapper[4980]: I1206 03:36:33.982086 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b2da6efd-2b6d-4ca7-b056-32c4b132becf-config\") pod \"service-ca-operator-777779d784-r47r5\" (UID: \"b2da6efd-2b6d-4ca7-b056-32c4b132becf\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-r47r5" Dec 06 03:36:33 crc kubenswrapper[4980]: I1206 03:36:33.983828 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/46290605-e114-4beb-b1c9-b0389f20f5f3-trusted-ca\") pod \"ingress-operator-5b745b69d9-27gc7\" (UID: \"46290605-e114-4beb-b1c9-b0389f20f5f3\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-27gc7" Dec 06 03:36:33 crc kubenswrapper[4980]: I1206 03:36:33.994331 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/9b9bddd5-f192-4b4c-81ee-b523bf9c8705-plugins-dir\") pod \"csi-hostpathplugin-74hx2\" (UID: \"9b9bddd5-f192-4b4c-81ee-b523bf9c8705\") " pod="hostpath-provisioner/csi-hostpathplugin-74hx2" Dec 06 03:36:33 crc kubenswrapper[4980]: I1206 03:36:33.998941 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/9b9bddd5-f192-4b4c-81ee-b523bf9c8705-socket-dir\") pod \"csi-hostpathplugin-74hx2\" (UID: \"9b9bddd5-f192-4b4c-81ee-b523bf9c8705\") " pod="hostpath-provisioner/csi-hostpathplugin-74hx2" Dec 06 03:36:33 crc kubenswrapper[4980]: I1206 03:36:33.999023 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/9b9bddd5-f192-4b4c-81ee-b523bf9c8705-registration-dir\") pod \"csi-hostpathplugin-74hx2\" (UID: \"9b9bddd5-f192-4b4c-81ee-b523bf9c8705\") " pod="hostpath-provisioner/csi-hostpathplugin-74hx2" Dec 06 03:36:34 crc kubenswrapper[4980]: I1206 03:36:34.001206 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/05aaf766-565d-4b05-97ec-6c8078d33f85-images\") pod \"machine-config-operator-74547568cd-49749\" (UID: \"05aaf766-565d-4b05-97ec-6c8078d33f85\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-49749" Dec 06 03:36:34 crc kubenswrapper[4980]: I1206 03:36:34.007806 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/05aaf766-565d-4b05-97ec-6c8078d33f85-auth-proxy-config\") pod \"machine-config-operator-74547568cd-49749\" (UID: \"05aaf766-565d-4b05-97ec-6c8078d33f85\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-49749" Dec 06 03:36:34 crc kubenswrapper[4980]: I1206 03:36:34.008278 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f2555ff9-efdb-449a-9f88-67a3c80ad0cf-config\") pod \"kube-controller-manager-operator-78b949d7b-jp25g\" (UID: \"f2555ff9-efdb-449a-9f88-67a3c80ad0cf\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-jp25g" Dec 06 03:36:34 crc kubenswrapper[4980]: I1206 03:36:34.008408 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/3ee9a5b6-d84f-46d7-b3c2-aab85c634bd4-cert\") pod \"ingress-canary-799j5\" (UID: \"3ee9a5b6-d84f-46d7-b3c2-aab85c634bd4\") " pod="openshift-ingress-canary/ingress-canary-799j5" Dec 06 03:36:34 crc kubenswrapper[4980]: I1206 03:36:34.008615 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bw764\" (UniqueName: \"kubernetes.io/projected/768da7ce-a217-4749-be5b-c1ae2dd4e275-kube-api-access-bw764\") pod \"openshift-controller-manager-operator-756b6f6bc6-lrnhm\" (UID: \"768da7ce-a217-4749-be5b-c1ae2dd4e275\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-lrnhm" Dec 06 03:36:34 crc kubenswrapper[4980]: I1206 03:36:34.008717 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kzm64\" (UniqueName: \"kubernetes.io/projected/a42aad68-42cf-4264-9c94-8678b0eea1b8-kube-api-access-kzm64\") pod \"oauth-openshift-558db77b4-pmmcw\" (UID: \"a42aad68-42cf-4264-9c94-8678b0eea1b8\") " pod="openshift-authentication/oauth-openshift-558db77b4-pmmcw" Dec 06 03:36:34 crc kubenswrapper[4980]: I1206 03:36:34.008902 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c7146eb0-acc5-4bac-b429-e1e00761e666-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-8jc8f\" (UID: \"c7146eb0-acc5-4bac-b429-e1e00761e666\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-8jc8f" Dec 06 03:36:34 crc kubenswrapper[4980]: I1206 03:36:34.009154 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/acb53ac6-3f09-424c-a830-a1e15eec0d62-default-certificate\") pod \"router-default-5444994796-pzzkd\" (UID: \"acb53ac6-3f09-424c-a830-a1e15eec0d62\") " pod="openshift-ingress/router-default-5444994796-pzzkd" Dec 06 03:36:34 crc kubenswrapper[4980]: I1206 03:36:34.009303 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ccd9c9f8-3473-4f3d-884c-e545a0a502af-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-cjvdn\" (UID: \"ccd9c9f8-3473-4f3d-884c-e545a0a502af\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-cjvdn" Dec 06 03:36:34 crc kubenswrapper[4980]: I1206 03:36:34.009380 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/92b5e190-2712-40a7-a551-472805f3c84e-node-bootstrap-token\") pod \"machine-config-server-bfp7v\" (UID: \"92b5e190-2712-40a7-a551-472805f3c84e\") " pod="openshift-machine-config-operator/machine-config-server-bfp7v" Dec 06 03:36:34 crc kubenswrapper[4980]: I1206 03:36:34.009488 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/9b9bddd5-f192-4b4c-81ee-b523bf9c8705-csi-data-dir\") pod \"csi-hostpathplugin-74hx2\" (UID: \"9b9bddd5-f192-4b4c-81ee-b523bf9c8705\") " pod="hostpath-provisioner/csi-hostpathplugin-74hx2" Dec 06 03:36:34 crc kubenswrapper[4980]: I1206 03:36:34.009699 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/949f499a-5680-4fec-ab59-5ced4f24ba2f-webhook-cert\") pod \"packageserver-d55dfcdfc-gjdsq\" (UID: \"949f499a-5680-4fec-ab59-5ced4f24ba2f\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-gjdsq" Dec 06 03:36:34 crc kubenswrapper[4980]: I1206 03:36:34.009730 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/46290605-e114-4beb-b1c9-b0389f20f5f3-metrics-tls\") pod \"ingress-operator-5b745b69d9-27gc7\" (UID: \"46290605-e114-4beb-b1c9-b0389f20f5f3\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-27gc7" Dec 06 03:36:34 crc kubenswrapper[4980]: I1206 03:36:34.010023 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/8be432d7-9e4b-4452-81df-91998326f92b-metrics-tls\") pod \"dns-default-4dcjh\" (UID: \"8be432d7-9e4b-4452-81df-91998326f92b\") " pod="openshift-dns/dns-default-4dcjh" Dec 06 03:36:34 crc kubenswrapper[4980]: I1206 03:36:34.010012 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/9476a6a0-4d40-4c85-9bd1-2162c971e3dd-srv-cert\") pod \"olm-operator-6b444d44fb-sq6bp\" (UID: \"9476a6a0-4d40-4c85-9bd1-2162c971e3dd\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-sq6bp" Dec 06 03:36:34 crc kubenswrapper[4980]: I1206 03:36:34.011412 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/acb53ac6-3f09-424c-a830-a1e15eec0d62-service-ca-bundle\") pod \"router-default-5444994796-pzzkd\" (UID: \"acb53ac6-3f09-424c-a830-a1e15eec0d62\") " pod="openshift-ingress/router-default-5444994796-pzzkd" Dec 06 03:36:34 crc kubenswrapper[4980]: I1206 03:36:34.012991 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/2692cb89-853a-4c4e-a6a6-570673db6f14-signing-key\") pod \"service-ca-9c57cc56f-qpf6b\" (UID: \"2692cb89-853a-4c4e-a6a6-570673db6f14\") " pod="openshift-service-ca/service-ca-9c57cc56f-qpf6b" Dec 06 03:36:34 crc kubenswrapper[4980]: I1206 03:36:34.013478 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-257vm\" (UniqueName: \"kubernetes.io/projected/dae23bd6-4c2b-441e-b44b-b6977bbd974b-kube-api-access-257vm\") pod \"dns-operator-744455d44c-w6bmg\" (UID: \"dae23bd6-4c2b-441e-b44b-b6977bbd974b\") " pod="openshift-dns-operator/dns-operator-744455d44c-w6bmg" Dec 06 03:36:34 crc kubenswrapper[4980]: I1206 03:36:34.014089 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/acb53ac6-3f09-424c-a830-a1e15eec0d62-stats-auth\") pod \"router-default-5444994796-pzzkd\" (UID: \"acb53ac6-3f09-424c-a830-a1e15eec0d62\") " pod="openshift-ingress/router-default-5444994796-pzzkd" Dec 06 03:36:34 crc kubenswrapper[4980]: I1206 03:36:34.014551 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/05aaf766-565d-4b05-97ec-6c8078d33f85-proxy-tls\") pod \"machine-config-operator-74547568cd-49749\" (UID: \"05aaf766-565d-4b05-97ec-6c8078d33f85\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-49749" Dec 06 03:36:34 crc kubenswrapper[4980]: I1206 03:36:34.014759 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/03704eb9-227d-4985-87cf-acd66ae76037-secret-volume\") pod \"collect-profiles-29416530-9hm9m\" (UID: \"03704eb9-227d-4985-87cf-acd66ae76037\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416530-9hm9m" Dec 06 03:36:34 crc kubenswrapper[4980]: I1206 03:36:34.015823 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/8be432d7-9e4b-4452-81df-91998326f92b-config-volume\") pod \"dns-default-4dcjh\" (UID: \"8be432d7-9e4b-4452-81df-91998326f92b\") " pod="openshift-dns/dns-default-4dcjh" Dec 06 03:36:34 crc kubenswrapper[4980]: I1206 03:36:34.016471 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-pmmcw" Dec 06 03:36:34 crc kubenswrapper[4980]: I1206 03:36:34.019258 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/e1e4e0b4-0bfe-4b01-9c4e-710ba0ec03bd-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-qcn9f\" (UID: \"e1e4e0b4-0bfe-4b01-9c4e-710ba0ec03bd\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-qcn9f" Dec 06 03:36:34 crc kubenswrapper[4980]: I1206 03:36:34.019538 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/9476a6a0-4d40-4c85-9bd1-2162c971e3dd-profile-collector-cert\") pod \"olm-operator-6b444d44fb-sq6bp\" (UID: \"9476a6a0-4d40-4c85-9bd1-2162c971e3dd\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-sq6bp" Dec 06 03:36:34 crc kubenswrapper[4980]: I1206 03:36:34.025257 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-s6h9g" Dec 06 03:36:34 crc kubenswrapper[4980]: I1206 03:36:34.027286 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/949f499a-5680-4fec-ab59-5ced4f24ba2f-apiservice-cert\") pod \"packageserver-d55dfcdfc-gjdsq\" (UID: \"949f499a-5680-4fec-ab59-5ced4f24ba2f\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-gjdsq" Dec 06 03:36:34 crc kubenswrapper[4980]: I1206 03:36:34.028311 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bq8xx\" (UniqueName: \"kubernetes.io/projected/a3673c8a-c303-42ac-8565-f03e5d457aa1-kube-api-access-bq8xx\") pod \"authentication-operator-69f744f599-84ft2\" (UID: \"a3673c8a-c303-42ac-8565-f03e5d457aa1\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-84ft2" Dec 06 03:36:34 crc kubenswrapper[4980]: I1206 03:36:34.030429 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/d1d006b0-92b0-45d5-9c28-fd43f879b310-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-tlpd8\" (UID: \"d1d006b0-92b0-45d5-9c28-fd43f879b310\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-tlpd8" Dec 06 03:36:34 crc kubenswrapper[4980]: I1206 03:36:34.035382 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/2692cb89-853a-4c4e-a6a6-570673db6f14-signing-cabundle\") pod \"service-ca-9c57cc56f-qpf6b\" (UID: \"2692cb89-853a-4c4e-a6a6-570673db6f14\") " pod="openshift-service-ca/service-ca-9c57cc56f-qpf6b" Dec 06 03:36:34 crc kubenswrapper[4980]: I1206 03:36:34.038329 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/abfe3c42-9229-486f-8c87-8b883f6821e0-proxy-tls\") pod \"machine-config-controller-84d6567774-d6dd4\" (UID: \"abfe3c42-9229-486f-8c87-8b883f6821e0\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-d6dd4" Dec 06 03:36:34 crc kubenswrapper[4980]: I1206 03:36:34.038656 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/0a45bc54-be9c-4c4f-9cae-dc369d25be33-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-spd2v\" (UID: \"0a45bc54-be9c-4c4f-9cae-dc369d25be33\") " pod="openshift-marketplace/marketplace-operator-79b997595-spd2v" Dec 06 03:36:34 crc kubenswrapper[4980]: I1206 03:36:34.038953 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/a7ee597a-a4fc-45ef-841f-1d4f666ccfcf-profile-collector-cert\") pod \"catalog-operator-68c6474976-hjrkf\" (UID: \"a7ee597a-a4fc-45ef-841f-1d4f666ccfcf\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-hjrkf" Dec 06 03:36:34 crc kubenswrapper[4980]: I1206 03:36:34.041638 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"certs\" (UniqueName: \"kubernetes.io/secret/92b5e190-2712-40a7-a551-472805f3c84e-certs\") pod \"machine-config-server-bfp7v\" (UID: \"92b5e190-2712-40a7-a551-472805f3c84e\") " pod="openshift-machine-config-operator/machine-config-server-bfp7v" Dec 06 03:36:34 crc kubenswrapper[4980]: I1206 03:36:34.043333 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/a7ee597a-a4fc-45ef-841f-1d4f666ccfcf-srv-cert\") pod \"catalog-operator-68c6474976-hjrkf\" (UID: \"a7ee597a-a4fc-45ef-841f-1d4f666ccfcf\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-hjrkf" Dec 06 03:36:34 crc kubenswrapper[4980]: I1206 03:36:34.046570 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8kssv\" (UniqueName: \"kubernetes.io/projected/346bb011-c34e-4286-90d3-0242205e8e94-kube-api-access-8kssv\") pod \"downloads-7954f5f757-6h96g\" (UID: \"346bb011-c34e-4286-90d3-0242205e8e94\") " pod="openshift-console/downloads-7954f5f757-6h96g" Dec 06 03:36:34 crc kubenswrapper[4980]: I1206 03:36:34.049743 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f08eff3e-89fe-4caf-9ec1-b32ee99f3c10-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-xjtnv\" (UID: \"f08eff3e-89fe-4caf-9ec1-b32ee99f3c10\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-xjtnv" Dec 06 03:36:34 crc kubenswrapper[4980]: I1206 03:36:34.050488 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/e81d9a6d-991f-46b0-aba5-6f76a5b1031a-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-cqk6j\" (UID: \"e81d9a6d-991f-46b0-aba5-6f76a5b1031a\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-cqk6j" Dec 06 03:36:34 crc kubenswrapper[4980]: I1206 03:36:34.053186 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/b25ba819-6e4e-4ca0-a570-c2726b8fbec5-bound-sa-token\") pod \"image-registry-697d97f7c8-qw2fb\" (UID: \"b25ba819-6e4e-4ca0-a570-c2726b8fbec5\") " pod="openshift-image-registry/image-registry-697d97f7c8-qw2fb" Dec 06 03:36:34 crc kubenswrapper[4980]: I1206 03:36:34.061840 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-lrnhm" Dec 06 03:36:34 crc kubenswrapper[4980]: I1206 03:36:34.071089 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-w6bmg" Dec 06 03:36:34 crc kubenswrapper[4980]: I1206 03:36:34.071621 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 03:36:34 crc kubenswrapper[4980]: E1206 03:36:34.072304 4980 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 03:36:34.572269571 +0000 UTC m=+153.817020842 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 03:36:34 crc kubenswrapper[4980]: I1206 03:36:34.079752 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-kftgf" Dec 06 03:36:34 crc kubenswrapper[4980]: I1206 03:36:34.173313 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qw2fb\" (UID: \"b25ba819-6e4e-4ca0-a570-c2726b8fbec5\") " pod="openshift-image-registry/image-registry-697d97f7c8-qw2fb" Dec 06 03:36:34 crc kubenswrapper[4980]: E1206 03:36:34.174704 4980 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 03:36:34.674686308 +0000 UTC m=+153.919437579 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qw2fb" (UID: "b25ba819-6e4e-4ca0-a570-c2726b8fbec5") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 03:36:34 crc kubenswrapper[4980]: I1206 03:36:34.180927 4980 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-qpchc"] Dec 06 03:36:34 crc kubenswrapper[4980]: I1206 03:36:34.182589 4980 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-5zm4c"] Dec 06 03:36:34 crc kubenswrapper[4980]: I1206 03:36:34.200062 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bzls8\" (UniqueName: \"kubernetes.io/projected/b25ba819-6e4e-4ca0-a570-c2726b8fbec5-kube-api-access-bzls8\") pod \"image-registry-697d97f7c8-qw2fb\" (UID: \"b25ba819-6e4e-4ca0-a570-c2726b8fbec5\") " pod="openshift-image-registry/image-registry-697d97f7c8-qw2fb" Dec 06 03:36:34 crc kubenswrapper[4980]: I1206 03:36:34.201536 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wr5kj\" (UniqueName: \"kubernetes.io/projected/05aaf766-565d-4b05-97ec-6c8078d33f85-kube-api-access-wr5kj\") pod \"machine-config-operator-74547568cd-49749\" (UID: \"05aaf766-565d-4b05-97ec-6c8078d33f85\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-49749" Dec 06 03:36:34 crc kubenswrapper[4980]: I1206 03:36:34.201747 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ckcls\" (UniqueName: \"kubernetes.io/projected/acb53ac6-3f09-424c-a830-a1e15eec0d62-kube-api-access-ckcls\") pod \"router-default-5444994796-pzzkd\" (UID: \"acb53ac6-3f09-424c-a830-a1e15eec0d62\") " pod="openshift-ingress/router-default-5444994796-pzzkd" Dec 06 03:36:34 crc kubenswrapper[4980]: I1206 03:36:34.202082 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9p5zn\" (UniqueName: \"kubernetes.io/projected/46290605-e114-4beb-b1c9-b0389f20f5f3-kube-api-access-9p5zn\") pod \"ingress-operator-5b745b69d9-27gc7\" (UID: \"46290605-e114-4beb-b1c9-b0389f20f5f3\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-27gc7" Dec 06 03:36:34 crc kubenswrapper[4980]: I1206 03:36:34.204917 4980 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-x5m4p"] Dec 06 03:36:34 crc kubenswrapper[4980]: I1206 03:36:34.204961 4980 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-zqsvr"] Dec 06 03:36:34 crc kubenswrapper[4980]: I1206 03:36:34.206246 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x6x86\" (UniqueName: \"kubernetes.io/projected/a7ee597a-a4fc-45ef-841f-1d4f666ccfcf-kube-api-access-x6x86\") pod \"catalog-operator-68c6474976-hjrkf\" (UID: \"a7ee597a-a4fc-45ef-841f-1d4f666ccfcf\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-hjrkf" Dec 06 03:36:34 crc kubenswrapper[4980]: I1206 03:36:34.207321 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-hjrkf" Dec 06 03:36:34 crc kubenswrapper[4980]: I1206 03:36:34.207785 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t26fr\" (UniqueName: \"kubernetes.io/projected/b2da6efd-2b6d-4ca7-b056-32c4b132becf-kube-api-access-t26fr\") pod \"service-ca-operator-777779d784-r47r5\" (UID: \"b2da6efd-2b6d-4ca7-b056-32c4b132becf\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-r47r5" Dec 06 03:36:34 crc kubenswrapper[4980]: I1206 03:36:34.215645 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2w6nc\" (UniqueName: \"kubernetes.io/projected/03704eb9-227d-4985-87cf-acd66ae76037-kube-api-access-2w6nc\") pod \"collect-profiles-29416530-9hm9m\" (UID: \"03704eb9-227d-4985-87cf-acd66ae76037\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416530-9hm9m" Dec 06 03:36:34 crc kubenswrapper[4980]: I1206 03:36:34.215823 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/f2555ff9-efdb-449a-9f88-67a3c80ad0cf-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-jp25g\" (UID: \"f2555ff9-efdb-449a-9f88-67a3c80ad0cf\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-jp25g" Dec 06 03:36:34 crc kubenswrapper[4980]: I1206 03:36:34.235719 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-r47r5" Dec 06 03:36:34 crc kubenswrapper[4980]: I1206 03:36:34.236923 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z4m5n\" (UniqueName: \"kubernetes.io/projected/abfe3c42-9229-486f-8c87-8b883f6821e0-kube-api-access-z4m5n\") pod \"machine-config-controller-84d6567774-d6dd4\" (UID: \"abfe3c42-9229-486f-8c87-8b883f6821e0\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-d6dd4" Dec 06 03:36:34 crc kubenswrapper[4980]: W1206 03:36:34.240901 4980 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf8973f62_160b_4c4b_9238_6a2ff3d63bc5.slice/crio-2ef66b476aae38165c653d77c54f4634951ded49e17864eb30178861c6c89c99 WatchSource:0}: Error finding container 2ef66b476aae38165c653d77c54f4634951ded49e17864eb30178861c6c89c99: Status 404 returned error can't find the container with id 2ef66b476aae38165c653d77c54f4634951ded49e17864eb30178861c6c89c99 Dec 06 03:36:34 crc kubenswrapper[4980]: I1206 03:36:34.242820 4980 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-66bs5"] Dec 06 03:36:34 crc kubenswrapper[4980]: W1206 03:36:34.244245 4980 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7c3cb795_33ac_4b16_b41e_a7da0de03578.slice/crio-fc3aad662d3f4182070e19692a374bbd750444681aaa31ba212b04641cf71fab WatchSource:0}: Error finding container fc3aad662d3f4182070e19692a374bbd750444681aaa31ba212b04641cf71fab: Status 404 returned error can't find the container with id fc3aad662d3f4182070e19692a374bbd750444681aaa31ba212b04641cf71fab Dec 06 03:36:34 crc kubenswrapper[4980]: I1206 03:36:34.245321 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29416530-9hm9m" Dec 06 03:36:34 crc kubenswrapper[4980]: I1206 03:36:34.248950 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/c7146eb0-acc5-4bac-b429-e1e00761e666-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-8jc8f\" (UID: \"c7146eb0-acc5-4bac-b429-e1e00761e666\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-8jc8f" Dec 06 03:36:34 crc kubenswrapper[4980]: I1206 03:36:34.274293 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 03:36:34 crc kubenswrapper[4980]: E1206 03:36:34.274790 4980 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 03:36:34.774762767 +0000 UTC m=+154.019514038 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 03:36:34 crc kubenswrapper[4980]: I1206 03:36:34.277181 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vg8kf\" (UniqueName: \"kubernetes.io/projected/e81d9a6d-991f-46b0-aba5-6f76a5b1031a-kube-api-access-vg8kf\") pod \"control-plane-machine-set-operator-78cbb6b69f-cqk6j\" (UID: \"e81d9a6d-991f-46b0-aba5-6f76a5b1031a\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-cqk6j" Dec 06 03:36:34 crc kubenswrapper[4980]: I1206 03:36:34.285506 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-84ft2" Dec 06 03:36:34 crc kubenswrapper[4980]: I1206 03:36:34.290038 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pl4qc\" (UniqueName: \"kubernetes.io/projected/9b9bddd5-f192-4b4c-81ee-b523bf9c8705-kube-api-access-pl4qc\") pod \"csi-hostpathplugin-74hx2\" (UID: \"9b9bddd5-f192-4b4c-81ee-b523bf9c8705\") " pod="hostpath-provisioner/csi-hostpathplugin-74hx2" Dec 06 03:36:34 crc kubenswrapper[4980]: I1206 03:36:34.296691 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-6h96g" Dec 06 03:36:34 crc kubenswrapper[4980]: I1206 03:36:34.310266 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b6q5n\" (UniqueName: \"kubernetes.io/projected/8be432d7-9e4b-4452-81df-91998326f92b-kube-api-access-b6q5n\") pod \"dns-default-4dcjh\" (UID: \"8be432d7-9e4b-4452-81df-91998326f92b\") " pod="openshift-dns/dns-default-4dcjh" Dec 06 03:36:34 crc kubenswrapper[4980]: I1206 03:36:34.310923 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-74hx2" Dec 06 03:36:34 crc kubenswrapper[4980]: I1206 03:36:34.340026 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vcwpf\" (UniqueName: \"kubernetes.io/projected/9476a6a0-4d40-4c85-9bd1-2162c971e3dd-kube-api-access-vcwpf\") pod \"olm-operator-6b444d44fb-sq6bp\" (UID: \"9476a6a0-4d40-4c85-9bd1-2162c971e3dd\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-sq6bp" Dec 06 03:36:34 crc kubenswrapper[4980]: I1206 03:36:34.411751 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-pzzkd" Dec 06 03:36:34 crc kubenswrapper[4980]: I1206 03:36:34.415195 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qw2fb\" (UID: \"b25ba819-6e4e-4ca0-a570-c2726b8fbec5\") " pod="openshift-image-registry/image-registry-697d97f7c8-qw2fb" Dec 06 03:36:34 crc kubenswrapper[4980]: E1206 03:36:34.415577 4980 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 03:36:34.915566362 +0000 UTC m=+154.160317633 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qw2fb" (UID: "b25ba819-6e4e-4ca0-a570-c2726b8fbec5") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 03:36:34 crc kubenswrapper[4980]: I1206 03:36:34.415909 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-8jc8f" Dec 06 03:36:34 crc kubenswrapper[4980]: I1206 03:36:34.430826 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-49749" Dec 06 03:36:34 crc kubenswrapper[4980]: I1206 03:36:34.433756 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-d6dd4" Dec 06 03:36:34 crc kubenswrapper[4980]: I1206 03:36:34.438784 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n2vnn\" (UniqueName: \"kubernetes.io/projected/d1d006b0-92b0-45d5-9c28-fd43f879b310-kube-api-access-n2vnn\") pod \"multus-admission-controller-857f4d67dd-tlpd8\" (UID: \"d1d006b0-92b0-45d5-9c28-fd43f879b310\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-tlpd8" Dec 06 03:36:34 crc kubenswrapper[4980]: I1206 03:36:34.442731 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-jp25g" Dec 06 03:36:34 crc kubenswrapper[4980]: I1206 03:36:34.443181 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zswsd\" (UniqueName: \"kubernetes.io/projected/92b5e190-2712-40a7-a551-472805f3c84e-kube-api-access-zswsd\") pod \"machine-config-server-bfp7v\" (UID: \"92b5e190-2712-40a7-a551-472805f3c84e\") " pod="openshift-machine-config-operator/machine-config-server-bfp7v" Dec 06 03:36:34 crc kubenswrapper[4980]: I1206 03:36:34.443353 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f5l5z\" (UniqueName: \"kubernetes.io/projected/0a45bc54-be9c-4c4f-9cae-dc369d25be33-kube-api-access-f5l5z\") pod \"marketplace-operator-79b997595-spd2v\" (UID: \"0a45bc54-be9c-4c4f-9cae-dc369d25be33\") " pod="openshift-marketplace/marketplace-operator-79b997595-spd2v" Dec 06 03:36:34 crc kubenswrapper[4980]: I1206 03:36:34.444001 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dp275\" (UniqueName: \"kubernetes.io/projected/ccd9c9f8-3473-4f3d-884c-e545a0a502af-kube-api-access-dp275\") pod \"kube-storage-version-migrator-operator-b67b599dd-cjvdn\" (UID: \"ccd9c9f8-3473-4f3d-884c-e545a0a502af\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-cjvdn" Dec 06 03:36:34 crc kubenswrapper[4980]: I1206 03:36:34.449907 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/f08eff3e-89fe-4caf-9ec1-b32ee99f3c10-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-xjtnv\" (UID: \"f08eff3e-89fe-4caf-9ec1-b32ee99f3c10\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-xjtnv" Dec 06 03:36:34 crc kubenswrapper[4980]: I1206 03:36:34.450646 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-sq6bp" Dec 06 03:36:34 crc kubenswrapper[4980]: I1206 03:36:34.454186 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mfkfg\" (UniqueName: \"kubernetes.io/projected/d7fb8e4e-88e7-4cc9-baf9-6e217813b28d-kube-api-access-mfkfg\") pod \"migrator-59844c95c7-56s8p\" (UID: \"d7fb8e4e-88e7-4cc9-baf9-6e217813b28d\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-56s8p" Dec 06 03:36:34 crc kubenswrapper[4980]: I1206 03:36:34.460366 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-cqk6j" Dec 06 03:36:34 crc kubenswrapper[4980]: I1206 03:36:34.469355 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-cjvdn" Dec 06 03:36:34 crc kubenswrapper[4980]: I1206 03:36:34.470307 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/46290605-e114-4beb-b1c9-b0389f20f5f3-bound-sa-token\") pod \"ingress-operator-5b745b69d9-27gc7\" (UID: \"46290605-e114-4beb-b1c9-b0389f20f5f3\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-27gc7" Dec 06 03:36:34 crc kubenswrapper[4980]: I1206 03:36:34.479347 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-27gc7" Dec 06 03:36:34 crc kubenswrapper[4980]: I1206 03:36:34.526906 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-spd2v" Dec 06 03:36:34 crc kubenswrapper[4980]: I1206 03:36:34.527985 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-56s8p" Dec 06 03:36:34 crc kubenswrapper[4980]: I1206 03:36:34.528040 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 03:36:34 crc kubenswrapper[4980]: I1206 03:36:34.528472 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-tlpd8" Dec 06 03:36:34 crc kubenswrapper[4980]: E1206 03:36:34.528687 4980 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 03:36:35.028655387 +0000 UTC m=+154.273406658 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 03:36:34 crc kubenswrapper[4980]: I1206 03:36:34.556743 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t52g4\" (UniqueName: \"kubernetes.io/projected/3ee9a5b6-d84f-46d7-b3c2-aab85c634bd4-kube-api-access-t52g4\") pod \"ingress-canary-799j5\" (UID: \"3ee9a5b6-d84f-46d7-b3c2-aab85c634bd4\") " pod="openshift-ingress-canary/ingress-canary-799j5" Dec 06 03:36:34 crc kubenswrapper[4980]: I1206 03:36:34.558732 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hzvwm\" (UniqueName: \"kubernetes.io/projected/949f499a-5680-4fec-ab59-5ced4f24ba2f-kube-api-access-hzvwm\") pod \"packageserver-d55dfcdfc-gjdsq\" (UID: \"949f499a-5680-4fec-ab59-5ced4f24ba2f\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-gjdsq" Dec 06 03:36:34 crc kubenswrapper[4980]: I1206 03:36:34.562290 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c75dl\" (UniqueName: \"kubernetes.io/projected/e1e4e0b4-0bfe-4b01-9c4e-710ba0ec03bd-kube-api-access-c75dl\") pod \"package-server-manager-789f6589d5-qcn9f\" (UID: \"e1e4e0b4-0bfe-4b01-9c4e-710ba0ec03bd\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-qcn9f" Dec 06 03:36:34 crc kubenswrapper[4980]: I1206 03:36:34.569100 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-gjdsq" Dec 06 03:36:34 crc kubenswrapper[4980]: I1206 03:36:34.577491 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-4dcjh" Dec 06 03:36:34 crc kubenswrapper[4980]: I1206 03:36:34.585069 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-799j5" Dec 06 03:36:34 crc kubenswrapper[4980]: I1206 03:36:34.586345 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wn452\" (UniqueName: \"kubernetes.io/projected/2692cb89-853a-4c4e-a6a6-570673db6f14-kube-api-access-wn452\") pod \"service-ca-9c57cc56f-qpf6b\" (UID: \"2692cb89-853a-4c4e-a6a6-570673db6f14\") " pod="openshift-service-ca/service-ca-9c57cc56f-qpf6b" Dec 06 03:36:34 crc kubenswrapper[4980]: I1206 03:36:34.606771 4980 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-p4gmt"] Dec 06 03:36:34 crc kubenswrapper[4980]: I1206 03:36:34.607933 4980 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-9fg2s"] Dec 06 03:36:34 crc kubenswrapper[4980]: I1206 03:36:34.613653 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-bfp7v" Dec 06 03:36:34 crc kubenswrapper[4980]: I1206 03:36:34.629124 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qw2fb\" (UID: \"b25ba819-6e4e-4ca0-a570-c2726b8fbec5\") " pod="openshift-image-registry/image-registry-697d97f7c8-qw2fb" Dec 06 03:36:34 crc kubenswrapper[4980]: E1206 03:36:34.629542 4980 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 03:36:35.129508268 +0000 UTC m=+154.374259539 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qw2fb" (UID: "b25ba819-6e4e-4ca0-a570-c2726b8fbec5") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 03:36:34 crc kubenswrapper[4980]: I1206 03:36:34.771837 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 03:36:34 crc kubenswrapper[4980]: I1206 03:36:34.772420 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-xjtnv" Dec 06 03:36:34 crc kubenswrapper[4980]: E1206 03:36:34.772544 4980 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 03:36:35.272501006 +0000 UTC m=+154.517252277 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 03:36:34 crc kubenswrapper[4980]: I1206 03:36:34.774207 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-zqsvr" event={"ID":"7c3cb795-33ac-4b16-b41e-a7da0de03578","Type":"ContainerStarted","Data":"fc3aad662d3f4182070e19692a374bbd750444681aaa31ba212b04641cf71fab"} Dec 06 03:36:34 crc kubenswrapper[4980]: I1206 03:36:34.776826 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-6txpx" event={"ID":"f2ea7661-b67f-4fbb-924a-943fe48995cb","Type":"ContainerStarted","Data":"a2c130a26aa07e6815ab93da9ae44ab4391f6bc51205327e758a494cd74ce33a"} Dec 06 03:36:34 crc kubenswrapper[4980]: I1206 03:36:34.779426 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-425jl" event={"ID":"005c8e70-2c2b-43a4-ae31-fd9013840824","Type":"ContainerStarted","Data":"58b9ee0e1248a37ca0518084fd0acf07490d43b138d8631c0bdc915bc20c7393"} Dec 06 03:36:34 crc kubenswrapper[4980]: I1206 03:36:34.782344 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-2ph5v" event={"ID":"c7db3f86-0e71-46d1-b95c-929069bf8b3a","Type":"ContainerStarted","Data":"2ef8b7cf1dcac906424d4e7209c65dd7daa71329776957ae1672c27fedffca30"} Dec 06 03:36:34 crc kubenswrapper[4980]: I1206 03:36:34.782378 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-2ph5v" event={"ID":"c7db3f86-0e71-46d1-b95c-929069bf8b3a","Type":"ContainerStarted","Data":"cef8e79f42a30641892997d0344f76c48e1de2b456ff814a4372bab18c9bbabb"} Dec 06 03:36:34 crc kubenswrapper[4980]: I1206 03:36:34.783079 4980 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-2ph5v" Dec 06 03:36:34 crc kubenswrapper[4980]: I1206 03:36:34.788393 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-qpchc" event={"ID":"2ac9a45e-ea4a-472e-bd5a-d947a0e2713e","Type":"ContainerStarted","Data":"612c0487e2671aaa03994e65c2de3a79a15add54047fd57677387ae7f740077f"} Dec 06 03:36:34 crc kubenswrapper[4980]: I1206 03:36:34.791060 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-x5m4p" event={"ID":"f8973f62-160b-4c4b-9238-6a2ff3d63bc5","Type":"ContainerStarted","Data":"2ef66b476aae38165c653d77c54f4634951ded49e17864eb30178861c6c89c99"} Dec 06 03:36:34 crc kubenswrapper[4980]: I1206 03:36:34.798227 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-qcn9f" Dec 06 03:36:34 crc kubenswrapper[4980]: I1206 03:36:34.803776 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-66bs5" event={"ID":"41587982-d7bb-413c-a1b4-12baa2ea600a","Type":"ContainerStarted","Data":"01231cb9d84da868a258f5f25f38e662b6917aee4be91cc487fc6cad130ff519"} Dec 06 03:36:34 crc kubenswrapper[4980]: I1206 03:36:34.805980 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-gjmj8" event={"ID":"443a2dc9-0546-4ab0-a448-eab81c09b482","Type":"ContainerStarted","Data":"f9b288fe9c96084f659c67bb671e2f5829adc803876882688aa7d328612221d3"} Dec 06 03:36:34 crc kubenswrapper[4980]: I1206 03:36:34.806009 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-gjmj8" event={"ID":"443a2dc9-0546-4ab0-a448-eab81c09b482","Type":"ContainerStarted","Data":"910dc6874f9328f4b53503bac76edcaf8fb09966143b238c1d0811e992afbc15"} Dec 06 03:36:34 crc kubenswrapper[4980]: I1206 03:36:34.856991 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-qpf6b" Dec 06 03:36:34 crc kubenswrapper[4980]: I1206 03:36:34.873991 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qw2fb\" (UID: \"b25ba819-6e4e-4ca0-a570-c2726b8fbec5\") " pod="openshift-image-registry/image-registry-697d97f7c8-qw2fb" Dec 06 03:36:34 crc kubenswrapper[4980]: E1206 03:36:34.876355 4980 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 03:36:35.376339924 +0000 UTC m=+154.621091225 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qw2fb" (UID: "b25ba819-6e4e-4ca0-a570-c2726b8fbec5") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 03:36:34 crc kubenswrapper[4980]: I1206 03:36:34.957320 4980 patch_prober.go:28] interesting pod/route-controller-manager-6576b87f9c-2ph5v container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.8:8443/healthz\": dial tcp 10.217.0.8:8443: connect: connection refused" start-of-body= Dec 06 03:36:34 crc kubenswrapper[4980]: I1206 03:36:34.957607 4980 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-2ph5v" podUID="c7db3f86-0e71-46d1-b95c-929069bf8b3a" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.8:8443/healthz\": dial tcp 10.217.0.8:8443: connect: connection refused" Dec 06 03:36:34 crc kubenswrapper[4980]: W1206 03:36:34.967486 4980 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podacb53ac6_3f09_424c_a830_a1e15eec0d62.slice/crio-0a1e0cc7c0e135ec12d2692f7e318162fc62f0daec839303d65b3443747e25a2 WatchSource:0}: Error finding container 0a1e0cc7c0e135ec12d2692f7e318162fc62f0daec839303d65b3443747e25a2: Status 404 returned error can't find the container with id 0a1e0cc7c0e135ec12d2692f7e318162fc62f0daec839303d65b3443747e25a2 Dec 06 03:36:34 crc kubenswrapper[4980]: I1206 03:36:34.979918 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 03:36:34 crc kubenswrapper[4980]: E1206 03:36:34.980354 4980 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 03:36:35.480329816 +0000 UTC m=+154.725081087 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 03:36:34 crc kubenswrapper[4980]: I1206 03:36:34.980461 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qw2fb\" (UID: \"b25ba819-6e4e-4ca0-a570-c2726b8fbec5\") " pod="openshift-image-registry/image-registry-697d97f7c8-qw2fb" Dec 06 03:36:34 crc kubenswrapper[4980]: E1206 03:36:34.981698 4980 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 03:36:35.481643594 +0000 UTC m=+154.726394865 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qw2fb" (UID: "b25ba819-6e4e-4ca0-a570-c2726b8fbec5") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 03:36:35 crc kubenswrapper[4980]: I1206 03:36:35.081024 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 03:36:35 crc kubenswrapper[4980]: E1206 03:36:35.095292 4980 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 03:36:35.595269734 +0000 UTC m=+154.840021005 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 03:36:35 crc kubenswrapper[4980]: I1206 03:36:35.196203 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qw2fb\" (UID: \"b25ba819-6e4e-4ca0-a570-c2726b8fbec5\") " pod="openshift-image-registry/image-registry-697d97f7c8-qw2fb" Dec 06 03:36:35 crc kubenswrapper[4980]: E1206 03:36:35.196709 4980 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 03:36:35.696694171 +0000 UTC m=+154.941445442 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qw2fb" (UID: "b25ba819-6e4e-4ca0-a570-c2726b8fbec5") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 03:36:35 crc kubenswrapper[4980]: I1206 03:36:35.321126 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 03:36:35 crc kubenswrapper[4980]: E1206 03:36:35.321873 4980 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 03:36:35.821854605 +0000 UTC m=+155.066605876 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 03:36:35 crc kubenswrapper[4980]: I1206 03:36:35.461888 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qw2fb\" (UID: \"b25ba819-6e4e-4ca0-a570-c2726b8fbec5\") " pod="openshift-image-registry/image-registry-697d97f7c8-qw2fb" Dec 06 03:36:35 crc kubenswrapper[4980]: E1206 03:36:35.462473 4980 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 03:36:35.962459824 +0000 UTC m=+155.207211095 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qw2fb" (UID: "b25ba819-6e4e-4ca0-a570-c2726b8fbec5") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 03:36:35 crc kubenswrapper[4980]: I1206 03:36:35.566327 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 03:36:35 crc kubenswrapper[4980]: E1206 03:36:35.566829 4980 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 03:36:36.066810057 +0000 UTC m=+155.311561328 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 03:36:35 crc kubenswrapper[4980]: I1206 03:36:35.667450 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qw2fb\" (UID: \"b25ba819-6e4e-4ca0-a570-c2726b8fbec5\") " pod="openshift-image-registry/image-registry-697d97f7c8-qw2fb" Dec 06 03:36:35 crc kubenswrapper[4980]: E1206 03:36:35.667850 4980 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 03:36:36.167838533 +0000 UTC m=+155.412589804 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qw2fb" (UID: "b25ba819-6e4e-4ca0-a570-c2726b8fbec5") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 03:36:35 crc kubenswrapper[4980]: I1206 03:36:35.824135 4980 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-2ph5v" podStartSLOduration=132.824100444 podStartE2EDuration="2m12.824100444s" podCreationTimestamp="2025-12-06 03:34:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 03:36:35.721455891 +0000 UTC m=+154.966207162" watchObservedRunningTime="2025-12-06 03:36:35.824100444 +0000 UTC m=+155.068851725" Dec 06 03:36:35 crc kubenswrapper[4980]: I1206 03:36:35.826933 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 03:36:35 crc kubenswrapper[4980]: E1206 03:36:35.827167 4980 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 03:36:36.327141352 +0000 UTC m=+155.571892643 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 03:36:35 crc kubenswrapper[4980]: I1206 03:36:35.858997 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-bfp7v" event={"ID":"92b5e190-2712-40a7-a551-472805f3c84e","Type":"ContainerStarted","Data":"16478f17b10776c3d1dee9636dbde0cb5e0ea2257ccad6be82b03cc30ac0cee7"} Dec 06 03:36:35 crc kubenswrapper[4980]: I1206 03:36:35.859053 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-bfp7v" event={"ID":"92b5e190-2712-40a7-a551-472805f3c84e","Type":"ContainerStarted","Data":"5525ab3c12e6a19f11ef825e75c8d708882d80cb3c442d1536a9b610435c247f"} Dec 06 03:36:35 crc kubenswrapper[4980]: I1206 03:36:35.874857 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-9fg2s" event={"ID":"a83bd6cc-62d8-44f4-930b-1c8fb70277c5","Type":"ContainerStarted","Data":"e0afb82445b15d3b250bc35f5b3b6b6ad078af8a0de753fb3f870becf2809264"} Dec 06 03:36:35 crc kubenswrapper[4980]: I1206 03:36:35.874913 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-9fg2s" event={"ID":"a83bd6cc-62d8-44f4-930b-1c8fb70277c5","Type":"ContainerStarted","Data":"96758a4ce394bd2dd972a372628707a401e1f6ee63cd387313aebf067c42b0aa"} Dec 06 03:36:35 crc kubenswrapper[4980]: I1206 03:36:35.892565 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-pzzkd" event={"ID":"acb53ac6-3f09-424c-a830-a1e15eec0d62","Type":"ContainerStarted","Data":"0a1e0cc7c0e135ec12d2692f7e318162fc62f0daec839303d65b3443747e25a2"} Dec 06 03:36:35 crc kubenswrapper[4980]: I1206 03:36:35.919880 4980 generic.go:334] "Generic (PLEG): container finished" podID="7c3cb795-33ac-4b16-b41e-a7da0de03578" containerID="a9703202f2a29c3de87f8e046b9ce580161b0b6227a9373ac75c3e2c081f085a" exitCode=0 Dec 06 03:36:35 crc kubenswrapper[4980]: I1206 03:36:35.919983 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-zqsvr" event={"ID":"7c3cb795-33ac-4b16-b41e-a7da0de03578","Type":"ContainerDied","Data":"a9703202f2a29c3de87f8e046b9ce580161b0b6227a9373ac75c3e2c081f085a"} Dec 06 03:36:35 crc kubenswrapper[4980]: I1206 03:36:35.929443 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qw2fb\" (UID: \"b25ba819-6e4e-4ca0-a570-c2726b8fbec5\") " pod="openshift-image-registry/image-registry-697d97f7c8-qw2fb" Dec 06 03:36:35 crc kubenswrapper[4980]: E1206 03:36:35.929889 4980 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 03:36:36.429875638 +0000 UTC m=+155.674626919 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qw2fb" (UID: "b25ba819-6e4e-4ca0-a570-c2726b8fbec5") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 03:36:35 crc kubenswrapper[4980]: I1206 03:36:35.941735 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-p4gmt" event={"ID":"f3921e9a-9d9a-45e0-8733-6a1cd20422d2","Type":"ContainerStarted","Data":"bc4d8a45f8ab917212ed3358399e413288b42799bdb74a5822bd5ae64f22e706"} Dec 06 03:36:35 crc kubenswrapper[4980]: I1206 03:36:35.963904 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-qpchc" event={"ID":"2ac9a45e-ea4a-472e-bd5a-d947a0e2713e","Type":"ContainerStarted","Data":"288edbab674e96c7efebe2535fb2456297f6f828a2250c216e9aef3578076bff"} Dec 06 03:36:35 crc kubenswrapper[4980]: I1206 03:36:35.965221 4980 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-879f6c89f-qpchc" Dec 06 03:36:35 crc kubenswrapper[4980]: I1206 03:36:35.978288 4980 patch_prober.go:28] interesting pod/controller-manager-879f6c89f-qpchc container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.7:8443/healthz\": dial tcp 10.217.0.7:8443: connect: connection refused" start-of-body= Dec 06 03:36:35 crc kubenswrapper[4980]: I1206 03:36:35.978381 4980 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-879f6c89f-qpchc" podUID="2ac9a45e-ea4a-472e-bd5a-d947a0e2713e" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.7:8443/healthz\": dial tcp 10.217.0.7:8443: connect: connection refused" Dec 06 03:36:35 crc kubenswrapper[4980]: I1206 03:36:35.979123 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-5zm4c" event={"ID":"03939884-e3c9-47b9-a7cb-bc19e2f09c9c","Type":"ContainerStarted","Data":"1d1a929b2a4002ac14d5d6a3175dc6dcf85a72cf421dd5cf1dc1ade879503c68"} Dec 06 03:36:35 crc kubenswrapper[4980]: I1206 03:36:35.979201 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-5zm4c" event={"ID":"03939884-e3c9-47b9-a7cb-bc19e2f09c9c","Type":"ContainerStarted","Data":"50d50449467018504cf3d9b325aaa10205a228ebcdb5bae73a81ad5488d8da41"} Dec 06 03:36:36 crc kubenswrapper[4980]: I1206 03:36:36.002490 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-x5m4p" event={"ID":"f8973f62-160b-4c4b-9238-6a2ff3d63bc5","Type":"ContainerStarted","Data":"5240df85fcdeb3418fd8814f57671cdfa6d36851156b9d6f33dd31b9094468be"} Dec 06 03:36:36 crc kubenswrapper[4980]: I1206 03:36:36.030617 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 03:36:36 crc kubenswrapper[4980]: E1206 03:36:36.032211 4980 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 03:36:36.532181152 +0000 UTC m=+155.776932493 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 03:36:36 crc kubenswrapper[4980]: I1206 03:36:36.132412 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qw2fb\" (UID: \"b25ba819-6e4e-4ca0-a570-c2726b8fbec5\") " pod="openshift-image-registry/image-registry-697d97f7c8-qw2fb" Dec 06 03:36:36 crc kubenswrapper[4980]: E1206 03:36:36.135364 4980 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 03:36:36.63534754 +0000 UTC m=+155.880098811 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qw2fb" (UID: "b25ba819-6e4e-4ca0-a570-c2726b8fbec5") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 03:36:36 crc kubenswrapper[4980]: I1206 03:36:36.139901 4980 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-425jl" podStartSLOduration=135.139875921 podStartE2EDuration="2m15.139875921s" podCreationTimestamp="2025-12-06 03:34:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 03:36:36.128185943 +0000 UTC m=+155.372937224" watchObservedRunningTime="2025-12-06 03:36:36.139875921 +0000 UTC m=+155.384627212" Dec 06 03:36:36 crc kubenswrapper[4980]: I1206 03:36:36.234472 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 03:36:36 crc kubenswrapper[4980]: E1206 03:36:36.234985 4980 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 03:36:36.734965426 +0000 UTC m=+155.979716697 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 03:36:36 crc kubenswrapper[4980]: I1206 03:36:36.336033 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qw2fb\" (UID: \"b25ba819-6e4e-4ca0-a570-c2726b8fbec5\") " pod="openshift-image-registry/image-registry-697d97f7c8-qw2fb" Dec 06 03:36:36 crc kubenswrapper[4980]: E1206 03:36:36.336476 4980 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 03:36:36.836460306 +0000 UTC m=+156.081211577 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qw2fb" (UID: "b25ba819-6e4e-4ca0-a570-c2726b8fbec5") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 03:36:36 crc kubenswrapper[4980]: I1206 03:36:36.404966 4980 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-2ph5v" Dec 06 03:36:36 crc kubenswrapper[4980]: I1206 03:36:36.418003 4980 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-6txpx" podStartSLOduration=134.417978139 podStartE2EDuration="2m14.417978139s" podCreationTimestamp="2025-12-06 03:34:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 03:36:36.416779965 +0000 UTC m=+155.661531236" watchObservedRunningTime="2025-12-06 03:36:36.417978139 +0000 UTC m=+155.662729410" Dec 06 03:36:36 crc kubenswrapper[4980]: I1206 03:36:36.437629 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 03:36:36 crc kubenswrapper[4980]: E1206 03:36:36.437789 4980 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 03:36:36.937768801 +0000 UTC m=+156.182520072 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 03:36:36 crc kubenswrapper[4980]: I1206 03:36:36.437889 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qw2fb\" (UID: \"b25ba819-6e4e-4ca0-a570-c2726b8fbec5\") " pod="openshift-image-registry/image-registry-697d97f7c8-qw2fb" Dec 06 03:36:36 crc kubenswrapper[4980]: E1206 03:36:36.438180 4980 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 03:36:36.938173152 +0000 UTC m=+156.182924423 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qw2fb" (UID: "b25ba819-6e4e-4ca0-a570-c2726b8fbec5") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 03:36:36 crc kubenswrapper[4980]: I1206 03:36:36.537805 4980 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-f9d7485db-gjmj8" podStartSLOduration=134.537789458 podStartE2EDuration="2m14.537789458s" podCreationTimestamp="2025-12-06 03:34:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 03:36:36.497381242 +0000 UTC m=+155.742132513" watchObservedRunningTime="2025-12-06 03:36:36.537789458 +0000 UTC m=+155.782540729" Dec 06 03:36:36 crc kubenswrapper[4980]: I1206 03:36:36.539682 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 03:36:36 crc kubenswrapper[4980]: E1206 03:36:36.539971 4980 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 03:36:37.039958221 +0000 UTC m=+156.284709492 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 03:36:36 crc kubenswrapper[4980]: I1206 03:36:36.624644 4980 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-server-bfp7v" podStartSLOduration=5.624627365 podStartE2EDuration="5.624627365s" podCreationTimestamp="2025-12-06 03:36:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 03:36:36.565394385 +0000 UTC m=+155.810145656" watchObservedRunningTime="2025-12-06 03:36:36.624627365 +0000 UTC m=+155.869378636" Dec 06 03:36:36 crc kubenswrapper[4980]: I1206 03:36:36.626106 4980 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-9fg2s" podStartSLOduration=133.626098638 podStartE2EDuration="2m13.626098638s" podCreationTimestamp="2025-12-06 03:34:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 03:36:36.622527195 +0000 UTC m=+155.867278466" watchObservedRunningTime="2025-12-06 03:36:36.626098638 +0000 UTC m=+155.870849909" Dec 06 03:36:36 crc kubenswrapper[4980]: I1206 03:36:36.654595 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qw2fb\" (UID: \"b25ba819-6e4e-4ca0-a570-c2726b8fbec5\") " pod="openshift-image-registry/image-registry-697d97f7c8-qw2fb" Dec 06 03:36:36 crc kubenswrapper[4980]: E1206 03:36:36.654970 4980 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 03:36:37.154953611 +0000 UTC m=+156.399704872 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qw2fb" (UID: "b25ba819-6e4e-4ca0-a570-c2726b8fbec5") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 03:36:36 crc kubenswrapper[4980]: I1206 03:36:36.683564 4980 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-879f6c89f-qpchc" podStartSLOduration=133.683543616 podStartE2EDuration="2m13.683543616s" podCreationTimestamp="2025-12-06 03:34:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 03:36:36.681640771 +0000 UTC m=+155.926392052" watchObservedRunningTime="2025-12-06 03:36:36.683543616 +0000 UTC m=+155.928294887" Dec 06 03:36:36 crc kubenswrapper[4980]: I1206 03:36:36.769044 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 03:36:36 crc kubenswrapper[4980]: E1206 03:36:36.769226 4980 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 03:36:37.269197789 +0000 UTC m=+156.513949060 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 03:36:36 crc kubenswrapper[4980]: I1206 03:36:36.769329 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qw2fb\" (UID: \"b25ba819-6e4e-4ca0-a570-c2726b8fbec5\") " pod="openshift-image-registry/image-registry-697d97f7c8-qw2fb" Dec 06 03:36:36 crc kubenswrapper[4980]: E1206 03:36:36.769655 4980 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 03:36:37.269647952 +0000 UTC m=+156.514399223 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qw2fb" (UID: "b25ba819-6e4e-4ca0-a570-c2726b8fbec5") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 03:36:36 crc kubenswrapper[4980]: I1206 03:36:36.869995 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 03:36:36 crc kubenswrapper[4980]: E1206 03:36:36.870316 4980 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 03:36:37.370268056 +0000 UTC m=+156.615019327 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 03:36:36 crc kubenswrapper[4980]: I1206 03:36:36.971539 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qw2fb\" (UID: \"b25ba819-6e4e-4ca0-a570-c2726b8fbec5\") " pod="openshift-image-registry/image-registry-697d97f7c8-qw2fb" Dec 06 03:36:36 crc kubenswrapper[4980]: E1206 03:36:36.971972 4980 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 03:36:37.471949652 +0000 UTC m=+156.716700973 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qw2fb" (UID: "b25ba819-6e4e-4ca0-a570-c2726b8fbec5") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 03:36:37 crc kubenswrapper[4980]: I1206 03:36:37.010038 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-x5m4p" event={"ID":"f8973f62-160b-4c4b-9238-6a2ff3d63bc5","Type":"ContainerStarted","Data":"04d8ac80b1b2cede8edbfdd572d1ceefe511f3ed71a91bcfe9b9b87522ac7718"} Dec 06 03:36:37 crc kubenswrapper[4980]: I1206 03:36:37.011782 4980 generic.go:334] "Generic (PLEG): container finished" podID="41587982-d7bb-413c-a1b4-12baa2ea600a" containerID="80defb56ac3b0f290288168c4807328315770a9734b96e29f5c14e8a898bfa4e" exitCode=0 Dec 06 03:36:37 crc kubenswrapper[4980]: I1206 03:36:37.011831 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-66bs5" event={"ID":"41587982-d7bb-413c-a1b4-12baa2ea600a","Type":"ContainerDied","Data":"80defb56ac3b0f290288168c4807328315770a9734b96e29f5c14e8a898bfa4e"} Dec 06 03:36:37 crc kubenswrapper[4980]: I1206 03:36:37.013325 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-pzzkd" event={"ID":"acb53ac6-3f09-424c-a830-a1e15eec0d62","Type":"ContainerStarted","Data":"95058d04f5d0a4d52317c5c118b7e797f7bee08cad36294b0bde1bde34493995"} Dec 06 03:36:37 crc kubenswrapper[4980]: I1206 03:36:37.016248 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-zqsvr" event={"ID":"7c3cb795-33ac-4b16-b41e-a7da0de03578","Type":"ContainerStarted","Data":"c5354dd2f408e95ba4da79df17fa95d50b8b96b1f3ddd4fbdfa74f546e975fed"} Dec 06 03:36:37 crc kubenswrapper[4980]: I1206 03:36:37.016500 4980 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-config-operator/openshift-config-operator-7777fb866f-zqsvr" Dec 06 03:36:37 crc kubenswrapper[4980]: I1206 03:36:37.017447 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-p4gmt" event={"ID":"f3921e9a-9d9a-45e0-8733-6a1cd20422d2","Type":"ContainerStarted","Data":"a78c47e711b1f065e268bbc396d742632e98c150b245ad9cdfc0c5d5df92112d"} Dec 06 03:36:37 crc kubenswrapper[4980]: I1206 03:36:37.019960 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-5zm4c" event={"ID":"03939884-e3c9-47b9-a7cb-bc19e2f09c9c","Type":"ContainerStarted","Data":"e0750ac61b3bc1e71644dba73d01f43ed1150cc856c7a1a2d7d964c1e7207a85"} Dec 06 03:36:37 crc kubenswrapper[4980]: I1206 03:36:37.022981 4980 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-879f6c89f-qpchc" Dec 06 03:36:37 crc kubenswrapper[4980]: I1206 03:36:37.072770 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 03:36:37 crc kubenswrapper[4980]: E1206 03:36:37.074448 4980 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 03:36:37.57442631 +0000 UTC m=+156.819177591 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 03:36:37 crc kubenswrapper[4980]: I1206 03:36:37.106541 4980 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/machine-api-operator-5694c8668f-x5m4p" podStartSLOduration=134.106490996 podStartE2EDuration="2m14.106490996s" podCreationTimestamp="2025-12-06 03:34:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 03:36:37.105777166 +0000 UTC m=+156.350528437" watchObservedRunningTime="2025-12-06 03:36:37.106490996 +0000 UTC m=+156.351242267" Dec 06 03:36:37 crc kubenswrapper[4980]: I1206 03:36:37.192766 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qw2fb\" (UID: \"b25ba819-6e4e-4ca0-a570-c2726b8fbec5\") " pod="openshift-image-registry/image-registry-697d97f7c8-qw2fb" Dec 06 03:36:37 crc kubenswrapper[4980]: E1206 03:36:37.193586 4980 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 03:36:37.69356076 +0000 UTC m=+156.938312031 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qw2fb" (UID: "b25ba819-6e4e-4ca0-a570-c2726b8fbec5") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 03:36:37 crc kubenswrapper[4980]: I1206 03:36:37.233422 4980 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd-operator/etcd-operator-b45778765-p4gmt" podStartSLOduration=134.23339598 podStartE2EDuration="2m14.23339598s" podCreationTimestamp="2025-12-06 03:34:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 03:36:37.227656864 +0000 UTC m=+156.472408135" watchObservedRunningTime="2025-12-06 03:36:37.23339598 +0000 UTC m=+156.478147251" Dec 06 03:36:37 crc kubenswrapper[4980]: I1206 03:36:37.255956 4980 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-kftgf"] Dec 06 03:36:37 crc kubenswrapper[4980]: I1206 03:36:37.293896 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 03:36:37 crc kubenswrapper[4980]: E1206 03:36:37.294315 4980 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 03:36:37.794294198 +0000 UTC m=+157.039045469 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 03:36:37 crc kubenswrapper[4980]: I1206 03:36:37.326184 4980 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-s6h9g"] Dec 06 03:36:37 crc kubenswrapper[4980]: I1206 03:36:37.333835 4980 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-config-operator/openshift-config-operator-7777fb866f-zqsvr" podStartSLOduration=135.333794278 podStartE2EDuration="2m15.333794278s" podCreationTimestamp="2025-12-06 03:34:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 03:36:37.278400629 +0000 UTC m=+156.523151910" watchObservedRunningTime="2025-12-06 03:36:37.333794278 +0000 UTC m=+156.578545539" Dec 06 03:36:37 crc kubenswrapper[4980]: I1206 03:36:37.371053 4980 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-lrnhm"] Dec 06 03:36:37 crc kubenswrapper[4980]: I1206 03:36:37.384313 4980 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-pmmcw"] Dec 06 03:36:37 crc kubenswrapper[4980]: I1206 03:36:37.388680 4980 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-5zm4c" podStartSLOduration=135.388655662 podStartE2EDuration="2m15.388655662s" podCreationTimestamp="2025-12-06 03:34:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 03:36:37.303639808 +0000 UTC m=+156.548391089" watchObservedRunningTime="2025-12-06 03:36:37.388655662 +0000 UTC m=+156.633406933" Dec 06 03:36:37 crc kubenswrapper[4980]: I1206 03:36:37.390426 4980 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress/router-default-5444994796-pzzkd" podStartSLOduration=134.390420063 podStartE2EDuration="2m14.390420063s" podCreationTimestamp="2025-12-06 03:34:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 03:36:37.334235861 +0000 UTC m=+156.578987132" watchObservedRunningTime="2025-12-06 03:36:37.390420063 +0000 UTC m=+156.635171334" Dec 06 03:36:37 crc kubenswrapper[4980]: I1206 03:36:37.396020 4980 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-w6bmg"] Dec 06 03:36:37 crc kubenswrapper[4980]: I1206 03:36:37.402422 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qw2fb\" (UID: \"b25ba819-6e4e-4ca0-a570-c2726b8fbec5\") " pod="openshift-image-registry/image-registry-697d97f7c8-qw2fb" Dec 06 03:36:37 crc kubenswrapper[4980]: E1206 03:36:37.403164 4980 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 03:36:37.903151021 +0000 UTC m=+157.147902292 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qw2fb" (UID: "b25ba819-6e4e-4ca0-a570-c2726b8fbec5") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 03:36:37 crc kubenswrapper[4980]: I1206 03:36:37.414848 4980 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-ingress/router-default-5444994796-pzzkd" Dec 06 03:36:37 crc kubenswrapper[4980]: I1206 03:36:37.422713 4980 patch_prober.go:28] interesting pod/router-default-5444994796-pzzkd container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 06 03:36:37 crc kubenswrapper[4980]: [-]has-synced failed: reason withheld Dec 06 03:36:37 crc kubenswrapper[4980]: [+]process-running ok Dec 06 03:36:37 crc kubenswrapper[4980]: healthz check failed Dec 06 03:36:37 crc kubenswrapper[4980]: I1206 03:36:37.422779 4980 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-pzzkd" podUID="acb53ac6-3f09-424c-a830-a1e15eec0d62" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 06 03:36:37 crc kubenswrapper[4980]: I1206 03:36:37.503379 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 03:36:37 crc kubenswrapper[4980]: E1206 03:36:37.503844 4980 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 03:36:38.003828237 +0000 UTC m=+157.248579508 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 03:36:37 crc kubenswrapper[4980]: I1206 03:36:37.662489 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qw2fb\" (UID: \"b25ba819-6e4e-4ca0-a570-c2726b8fbec5\") " pod="openshift-image-registry/image-registry-697d97f7c8-qw2fb" Dec 06 03:36:37 crc kubenswrapper[4980]: E1206 03:36:37.663243 4980 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 03:36:38.163226859 +0000 UTC m=+157.407978130 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qw2fb" (UID: "b25ba819-6e4e-4ca0-a570-c2726b8fbec5") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 03:36:37 crc kubenswrapper[4980]: I1206 03:36:37.764895 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 03:36:37 crc kubenswrapper[4980]: E1206 03:36:37.765311 4980 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 03:36:38.265295195 +0000 UTC m=+157.510046466 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 03:36:37 crc kubenswrapper[4980]: I1206 03:36:37.830195 4980 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-cqk6j"] Dec 06 03:36:37 crc kubenswrapper[4980]: I1206 03:36:37.855190 4980 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-799j5"] Dec 06 03:36:37 crc kubenswrapper[4980]: I1206 03:36:37.868069 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qw2fb\" (UID: \"b25ba819-6e4e-4ca0-a570-c2726b8fbec5\") " pod="openshift-image-registry/image-registry-697d97f7c8-qw2fb" Dec 06 03:36:37 crc kubenswrapper[4980]: E1206 03:36:37.868425 4980 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 03:36:38.368412102 +0000 UTC m=+157.613163363 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qw2fb" (UID: "b25ba819-6e4e-4ca0-a570-c2726b8fbec5") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 03:36:37 crc kubenswrapper[4980]: W1206 03:36:37.875659 4980 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3ee9a5b6_d84f_46d7_b3c2_aab85c634bd4.slice/crio-9247b3b89c29a3e4484a6d6a30c81336f3810059a9c5d3dbefe956c6d1514604 WatchSource:0}: Error finding container 9247b3b89c29a3e4484a6d6a30c81336f3810059a9c5d3dbefe956c6d1514604: Status 404 returned error can't find the container with id 9247b3b89c29a3e4484a6d6a30c81336f3810059a9c5d3dbefe956c6d1514604 Dec 06 03:36:37 crc kubenswrapper[4980]: I1206 03:36:37.902469 4980 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-8jc8f"] Dec 06 03:36:37 crc kubenswrapper[4980]: I1206 03:36:37.907086 4980 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-hjrkf"] Dec 06 03:36:37 crc kubenswrapper[4980]: I1206 03:36:37.915103 4980 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-6h96g"] Dec 06 03:36:37 crc kubenswrapper[4980]: I1206 03:36:37.926306 4980 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-xjtnv"] Dec 06 03:36:37 crc kubenswrapper[4980]: I1206 03:36:37.937292 4980 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29416530-9hm9m"] Dec 06 03:36:37 crc kubenswrapper[4980]: I1206 03:36:37.961400 4980 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-74hx2"] Dec 06 03:36:37 crc kubenswrapper[4980]: W1206 03:36:37.966195 4980 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod346bb011_c34e_4286_90d3_0242205e8e94.slice/crio-b941abdfaba56dd02747cb975cc8311347f981665e17b687381bb89b80c627c0 WatchSource:0}: Error finding container b941abdfaba56dd02747cb975cc8311347f981665e17b687381bb89b80c627c0: Status 404 returned error can't find the container with id b941abdfaba56dd02747cb975cc8311347f981665e17b687381bb89b80c627c0 Dec 06 03:36:37 crc kubenswrapper[4980]: I1206 03:36:37.969041 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 03:36:37 crc kubenswrapper[4980]: E1206 03:36:37.969414 4980 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 03:36:38.469400267 +0000 UTC m=+157.714151528 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 03:36:38 crc kubenswrapper[4980]: I1206 03:36:38.074939 4980 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-56s8p"] Dec 06 03:36:38 crc kubenswrapper[4980]: I1206 03:36:38.075242 4980 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-qpf6b"] Dec 06 03:36:38 crc kubenswrapper[4980]: I1206 03:36:38.077312 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qw2fb\" (UID: \"b25ba819-6e4e-4ca0-a570-c2726b8fbec5\") " pod="openshift-image-registry/image-registry-697d97f7c8-qw2fb" Dec 06 03:36:38 crc kubenswrapper[4980]: E1206 03:36:38.077595 4980 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 03:36:38.57758341 +0000 UTC m=+157.822334681 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qw2fb" (UID: "b25ba819-6e4e-4ca0-a570-c2726b8fbec5") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 03:36:38 crc kubenswrapper[4980]: I1206 03:36:38.094213 4980 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-sq6bp"] Dec 06 03:36:38 crc kubenswrapper[4980]: I1206 03:36:38.101444 4980 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-84ft2"] Dec 06 03:36:38 crc kubenswrapper[4980]: I1206 03:36:38.101501 4980 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-27gc7"] Dec 06 03:36:38 crc kubenswrapper[4980]: I1206 03:36:38.101534 4980 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-jp25g"] Dec 06 03:36:38 crc kubenswrapper[4980]: I1206 03:36:38.104371 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-6h96g" event={"ID":"346bb011-c34e-4286-90d3-0242205e8e94","Type":"ContainerStarted","Data":"b941abdfaba56dd02747cb975cc8311347f981665e17b687381bb89b80c627c0"} Dec 06 03:36:38 crc kubenswrapper[4980]: I1206 03:36:38.105434 4980 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-49749"] Dec 06 03:36:38 crc kubenswrapper[4980]: I1206 03:36:38.108213 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-8jc8f" event={"ID":"c7146eb0-acc5-4bac-b429-e1e00761e666","Type":"ContainerStarted","Data":"775cbb01b6c12baa936c037b4afe9c18e4755071716489a544230c258ecf8610"} Dec 06 03:36:38 crc kubenswrapper[4980]: I1206 03:36:38.153887 4980 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-r47r5"] Dec 06 03:36:38 crc kubenswrapper[4980]: I1206 03:36:38.215789 4980 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-qcn9f"] Dec 06 03:36:38 crc kubenswrapper[4980]: I1206 03:36:38.215878 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-w6bmg" event={"ID":"dae23bd6-4c2b-441e-b44b-b6977bbd974b","Type":"ContainerStarted","Data":"7219decbff839b6e33706450fe621da5e7d75cbc401fb904ffbb64b8759297ac"} Dec 06 03:36:38 crc kubenswrapper[4980]: I1206 03:36:38.215892 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 03:36:38 crc kubenswrapper[4980]: E1206 03:36:38.216314 4980 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 03:36:38.716297405 +0000 UTC m=+157.961048676 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 03:36:38 crc kubenswrapper[4980]: I1206 03:36:38.231110 4980 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-tlpd8"] Dec 06 03:36:38 crc kubenswrapper[4980]: I1206 03:36:38.234662 4980 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-d6dd4"] Dec 06 03:36:38 crc kubenswrapper[4980]: I1206 03:36:38.260896 4980 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-spd2v"] Dec 06 03:36:38 crc kubenswrapper[4980]: I1206 03:36:38.272572 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-cqk6j" event={"ID":"e81d9a6d-991f-46b0-aba5-6f76a5b1031a","Type":"ContainerStarted","Data":"cadd3be9d5619ba176180112fea493e2be79aae17e78d1cfe411edcccd0662d8"} Dec 06 03:36:38 crc kubenswrapper[4980]: I1206 03:36:38.280999 4980 generic.go:334] "Generic (PLEG): container finished" podID="2cf31abb-ef18-4ba5-8cc9-bf3f7507f771" containerID="98f76471637cfb4183f436d099ed45996a6efe2aa2b6665b82c184a9abad7947" exitCode=0 Dec 06 03:36:38 crc kubenswrapper[4980]: I1206 03:36:38.281072 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-s6h9g" event={"ID":"2cf31abb-ef18-4ba5-8cc9-bf3f7507f771","Type":"ContainerDied","Data":"98f76471637cfb4183f436d099ed45996a6efe2aa2b6665b82c184a9abad7947"} Dec 06 03:36:38 crc kubenswrapper[4980]: I1206 03:36:38.281107 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-s6h9g" event={"ID":"2cf31abb-ef18-4ba5-8cc9-bf3f7507f771","Type":"ContainerStarted","Data":"a8ed5e4efa93b2d21e3e52df95b993195c5a49300acf118dcd8ec296f4b75489"} Dec 06 03:36:38 crc kubenswrapper[4980]: I1206 03:36:38.288780 4980 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-cjvdn"] Dec 06 03:36:38 crc kubenswrapper[4980]: I1206 03:36:38.297061 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-hjrkf" event={"ID":"a7ee597a-a4fc-45ef-841f-1d4f666ccfcf","Type":"ContainerStarted","Data":"704a5676e5e77af400f2e785042666591cf17bca00b442e6d420354024db7090"} Dec 06 03:36:38 crc kubenswrapper[4980]: I1206 03:36:38.302594 4980 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-4dcjh"] Dec 06 03:36:38 crc kubenswrapper[4980]: I1206 03:36:38.317881 4980 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-gjdsq"] Dec 06 03:36:38 crc kubenswrapper[4980]: I1206 03:36:38.320375 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qw2fb\" (UID: \"b25ba819-6e4e-4ca0-a570-c2726b8fbec5\") " pod="openshift-image-registry/image-registry-697d97f7c8-qw2fb" Dec 06 03:36:38 crc kubenswrapper[4980]: E1206 03:36:38.320697 4980 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 03:36:38.820684449 +0000 UTC m=+158.065435720 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qw2fb" (UID: "b25ba819-6e4e-4ca0-a570-c2726b8fbec5") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 03:36:38 crc kubenswrapper[4980]: I1206 03:36:38.365357 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-66bs5" event={"ID":"41587982-d7bb-413c-a1b4-12baa2ea600a","Type":"ContainerStarted","Data":"6cb7853bd3f10b83be0720700d77993800e3ed83b48a090ac4a743f6f92af7dc"} Dec 06 03:36:38 crc kubenswrapper[4980]: I1206 03:36:38.418497 4980 patch_prober.go:28] interesting pod/router-default-5444994796-pzzkd container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 06 03:36:38 crc kubenswrapper[4980]: [-]has-synced failed: reason withheld Dec 06 03:36:38 crc kubenswrapper[4980]: [+]process-running ok Dec 06 03:36:38 crc kubenswrapper[4980]: healthz check failed Dec 06 03:36:38 crc kubenswrapper[4980]: I1206 03:36:38.418550 4980 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-pzzkd" podUID="acb53ac6-3f09-424c-a830-a1e15eec0d62" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 06 03:36:38 crc kubenswrapper[4980]: I1206 03:36:38.424759 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 03:36:38 crc kubenswrapper[4980]: E1206 03:36:38.425057 4980 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 03:36:38.925033441 +0000 UTC m=+158.169784702 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 03:36:38 crc kubenswrapper[4980]: I1206 03:36:38.425290 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qw2fb\" (UID: \"b25ba819-6e4e-4ca0-a570-c2726b8fbec5\") " pod="openshift-image-registry/image-registry-697d97f7c8-qw2fb" Dec 06 03:36:38 crc kubenswrapper[4980]: E1206 03:36:38.426701 4980 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 03:36:38.926687179 +0000 UTC m=+158.171438450 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qw2fb" (UID: "b25ba819-6e4e-4ca0-a570-c2726b8fbec5") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 03:36:38 crc kubenswrapper[4980]: I1206 03:36:38.427638 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-kftgf" event={"ID":"6d695618-ca31-48d0-a178-24d7ee34a54b","Type":"ContainerStarted","Data":"fa08b2d9f8d7e37c08b41816d14ff8124f545928eb4a69affd209fd7046fe7f9"} Dec 06 03:36:38 crc kubenswrapper[4980]: I1206 03:36:38.427676 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-kftgf" event={"ID":"6d695618-ca31-48d0-a178-24d7ee34a54b","Type":"ContainerStarted","Data":"6a0f1e03854011d61711df9feddd92d1728c07e60750d8ef1f993d6fd3185fdb"} Dec 06 03:36:38 crc kubenswrapper[4980]: I1206 03:36:38.428725 4980 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console-operator/console-operator-58897d9998-kftgf" Dec 06 03:36:38 crc kubenswrapper[4980]: I1206 03:36:38.447402 4980 patch_prober.go:28] interesting pod/console-operator-58897d9998-kftgf container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.32:8443/readyz\": dial tcp 10.217.0.32:8443: connect: connection refused" start-of-body= Dec 06 03:36:38 crc kubenswrapper[4980]: I1206 03:36:38.447470 4980 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-kftgf" podUID="6d695618-ca31-48d0-a178-24d7ee34a54b" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.32:8443/readyz\": dial tcp 10.217.0.32:8443: connect: connection refused" Dec 06 03:36:38 crc kubenswrapper[4980]: I1206 03:36:38.487079 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-pmmcw" event={"ID":"a42aad68-42cf-4264-9c94-8678b0eea1b8","Type":"ContainerStarted","Data":"679b95f34ba069aebe784949cf4ef7693988ae7acaa83b3bd44b881022728b30"} Dec 06 03:36:38 crc kubenswrapper[4980]: I1206 03:36:38.487231 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-pmmcw" event={"ID":"a42aad68-42cf-4264-9c94-8678b0eea1b8","Type":"ContainerStarted","Data":"95185ed7e417885129fd12417bd2f6923e4f65a41567a2a74c075efe7067b280"} Dec 06 03:36:38 crc kubenswrapper[4980]: I1206 03:36:38.489115 4980 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-558db77b4-pmmcw" Dec 06 03:36:38 crc kubenswrapper[4980]: I1206 03:36:38.494141 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-xjtnv" event={"ID":"f08eff3e-89fe-4caf-9ec1-b32ee99f3c10","Type":"ContainerStarted","Data":"c40d9f0a09189fbddf2e83406356ae63e86ece000ddbd066567ee57cd3416c1e"} Dec 06 03:36:38 crc kubenswrapper[4980]: I1206 03:36:38.498949 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-799j5" event={"ID":"3ee9a5b6-d84f-46d7-b3c2-aab85c634bd4","Type":"ContainerStarted","Data":"9247b3b89c29a3e4484a6d6a30c81336f3810059a9c5d3dbefe956c6d1514604"} Dec 06 03:36:38 crc kubenswrapper[4980]: I1206 03:36:38.505999 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-lrnhm" event={"ID":"768da7ce-a217-4749-be5b-c1ae2dd4e275","Type":"ContainerStarted","Data":"ebb595c8cea8e7193012995a121d2569625a63a527ea6134d3a3e5e5f2bcdcd5"} Dec 06 03:36:38 crc kubenswrapper[4980]: I1206 03:36:38.506050 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-lrnhm" event={"ID":"768da7ce-a217-4749-be5b-c1ae2dd4e275","Type":"ContainerStarted","Data":"e42ba5839e770401c1c9d906d9ee2180d4b7bef3da0d8bd75cc633f2ec36ca56"} Dec 06 03:36:38 crc kubenswrapper[4980]: I1206 03:36:38.525856 4980 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console-operator/console-operator-58897d9998-kftgf" podStartSLOduration=136.525836001 podStartE2EDuration="2m16.525836001s" podCreationTimestamp="2025-12-06 03:34:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 03:36:38.524238895 +0000 UTC m=+157.768990166" watchObservedRunningTime="2025-12-06 03:36:38.525836001 +0000 UTC m=+157.770587272" Dec 06 03:36:38 crc kubenswrapper[4980]: I1206 03:36:38.526564 4980 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-66bs5" podStartSLOduration=135.526557572 podStartE2EDuration="2m15.526557572s" podCreationTimestamp="2025-12-06 03:34:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 03:36:38.389036592 +0000 UTC m=+157.633787863" watchObservedRunningTime="2025-12-06 03:36:38.526557572 +0000 UTC m=+157.771308843" Dec 06 03:36:38 crc kubenswrapper[4980]: I1206 03:36:38.527883 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 03:36:38 crc kubenswrapper[4980]: E1206 03:36:38.528182 4980 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 03:36:39.028124187 +0000 UTC m=+158.272875458 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 03:36:38 crc kubenswrapper[4980]: I1206 03:36:38.528464 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qw2fb\" (UID: \"b25ba819-6e4e-4ca0-a570-c2726b8fbec5\") " pod="openshift-image-registry/image-registry-697d97f7c8-qw2fb" Dec 06 03:36:38 crc kubenswrapper[4980]: E1206 03:36:38.531098 4980 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 03:36:39.031080573 +0000 UTC m=+158.275831844 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qw2fb" (UID: "b25ba819-6e4e-4ca0-a570-c2726b8fbec5") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 03:36:38 crc kubenswrapper[4980]: I1206 03:36:38.626034 4980 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-lrnhm" podStartSLOduration=136.625997053 podStartE2EDuration="2m16.625997053s" podCreationTimestamp="2025-12-06 03:34:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 03:36:38.572854659 +0000 UTC m=+157.817605930" watchObservedRunningTime="2025-12-06 03:36:38.625997053 +0000 UTC m=+157.870748324" Dec 06 03:36:38 crc kubenswrapper[4980]: I1206 03:36:38.645099 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 03:36:38 crc kubenswrapper[4980]: E1206 03:36:38.645276 4980 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 03:36:39.145246989 +0000 UTC m=+158.389998260 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 03:36:38 crc kubenswrapper[4980]: I1206 03:36:38.645350 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qw2fb\" (UID: \"b25ba819-6e4e-4ca0-a570-c2726b8fbec5\") " pod="openshift-image-registry/image-registry-697d97f7c8-qw2fb" Dec 06 03:36:38 crc kubenswrapper[4980]: E1206 03:36:38.652112 4980 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 03:36:39.152090496 +0000 UTC m=+158.396841767 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qw2fb" (UID: "b25ba819-6e4e-4ca0-a570-c2726b8fbec5") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 03:36:38 crc kubenswrapper[4980]: I1206 03:36:38.703783 4980 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-canary/ingress-canary-799j5" podStartSLOduration=7.703756937 podStartE2EDuration="7.703756937s" podCreationTimestamp="2025-12-06 03:36:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 03:36:38.690944587 +0000 UTC m=+157.935695858" watchObservedRunningTime="2025-12-06 03:36:38.703756937 +0000 UTC m=+157.948508208" Dec 06 03:36:38 crc kubenswrapper[4980]: I1206 03:36:38.703956 4980 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-558db77b4-pmmcw" podStartSLOduration=136.703951282 podStartE2EDuration="2m16.703951282s" podCreationTimestamp="2025-12-06 03:34:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 03:36:38.634382565 +0000 UTC m=+157.879133836" watchObservedRunningTime="2025-12-06 03:36:38.703951282 +0000 UTC m=+157.948702553" Dec 06 03:36:38 crc kubenswrapper[4980]: I1206 03:36:38.762267 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 03:36:38 crc kubenswrapper[4980]: E1206 03:36:38.762931 4980 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 03:36:39.262909355 +0000 UTC m=+158.507660626 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 03:36:38 crc kubenswrapper[4980]: I1206 03:36:38.863600 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qw2fb\" (UID: \"b25ba819-6e4e-4ca0-a570-c2726b8fbec5\") " pod="openshift-image-registry/image-registry-697d97f7c8-qw2fb" Dec 06 03:36:38 crc kubenswrapper[4980]: E1206 03:36:38.864659 4980 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 03:36:39.364636121 +0000 UTC m=+158.609387402 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qw2fb" (UID: "b25ba819-6e4e-4ca0-a570-c2726b8fbec5") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 03:36:38 crc kubenswrapper[4980]: I1206 03:36:38.967146 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 03:36:38 crc kubenswrapper[4980]: E1206 03:36:38.967705 4980 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 03:36:39.467689656 +0000 UTC m=+158.712440927 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 03:36:39 crc kubenswrapper[4980]: I1206 03:36:39.068467 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qw2fb\" (UID: \"b25ba819-6e4e-4ca0-a570-c2726b8fbec5\") " pod="openshift-image-registry/image-registry-697d97f7c8-qw2fb" Dec 06 03:36:39 crc kubenswrapper[4980]: E1206 03:36:39.068785 4980 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 03:36:39.568773345 +0000 UTC m=+158.813524616 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qw2fb" (UID: "b25ba819-6e4e-4ca0-a570-c2726b8fbec5") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 03:36:39 crc kubenswrapper[4980]: I1206 03:36:39.169769 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 03:36:39 crc kubenswrapper[4980]: E1206 03:36:39.170011 4980 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 03:36:39.669954916 +0000 UTC m=+158.914706187 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 03:36:39 crc kubenswrapper[4980]: I1206 03:36:39.170111 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qw2fb\" (UID: \"b25ba819-6e4e-4ca0-a570-c2726b8fbec5\") " pod="openshift-image-registry/image-registry-697d97f7c8-qw2fb" Dec 06 03:36:39 crc kubenswrapper[4980]: E1206 03:36:39.170871 4980 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 03:36:39.670860992 +0000 UTC m=+158.915612263 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qw2fb" (UID: "b25ba819-6e4e-4ca0-a570-c2726b8fbec5") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 03:36:39 crc kubenswrapper[4980]: I1206 03:36:39.178557 4980 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-config-operator/openshift-config-operator-7777fb866f-zqsvr" Dec 06 03:36:39 crc kubenswrapper[4980]: I1206 03:36:39.270902 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 03:36:39 crc kubenswrapper[4980]: E1206 03:36:39.271681 4980 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 03:36:39.771655882 +0000 UTC m=+159.016407153 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 03:36:39 crc kubenswrapper[4980]: I1206 03:36:39.372029 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qw2fb\" (UID: \"b25ba819-6e4e-4ca0-a570-c2726b8fbec5\") " pod="openshift-image-registry/image-registry-697d97f7c8-qw2fb" Dec 06 03:36:39 crc kubenswrapper[4980]: E1206 03:36:39.372656 4980 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 03:36:39.872639937 +0000 UTC m=+159.117391208 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qw2fb" (UID: "b25ba819-6e4e-4ca0-a570-c2726b8fbec5") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 03:36:39 crc kubenswrapper[4980]: I1206 03:36:39.421253 4980 patch_prober.go:28] interesting pod/router-default-5444994796-pzzkd container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 06 03:36:39 crc kubenswrapper[4980]: [-]has-synced failed: reason withheld Dec 06 03:36:39 crc kubenswrapper[4980]: [+]process-running ok Dec 06 03:36:39 crc kubenswrapper[4980]: healthz check failed Dec 06 03:36:39 crc kubenswrapper[4980]: I1206 03:36:39.421314 4980 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-pzzkd" podUID="acb53ac6-3f09-424c-a830-a1e15eec0d62" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 06 03:36:39 crc kubenswrapper[4980]: I1206 03:36:39.496983 4980 patch_prober.go:28] interesting pod/oauth-openshift-558db77b4-pmmcw container/oauth-openshift namespace/openshift-authentication: Readiness probe status=failure output="Get \"https://10.217.0.25:6443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 06 03:36:39 crc kubenswrapper[4980]: I1206 03:36:39.497053 4980 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-authentication/oauth-openshift-558db77b4-pmmcw" podUID="a42aad68-42cf-4264-9c94-8678b0eea1b8" containerName="oauth-openshift" probeResult="failure" output="Get \"https://10.217.0.25:6443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 06 03:36:39 crc kubenswrapper[4980]: I1206 03:36:39.499817 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 03:36:39 crc kubenswrapper[4980]: E1206 03:36:39.500464 4980 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 03:36:40.000444087 +0000 UTC m=+159.245195358 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 03:36:39 crc kubenswrapper[4980]: I1206 03:36:39.604661 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-sq6bp" event={"ID":"9476a6a0-4d40-4c85-9bd1-2162c971e3dd","Type":"ContainerStarted","Data":"9719d5838c62abcbda88c4b69966450136090c3939291797e90f8686c3ec4e78"} Dec 06 03:36:39 crc kubenswrapper[4980]: I1206 03:36:39.611978 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qw2fb\" (UID: \"b25ba819-6e4e-4ca0-a570-c2726b8fbec5\") " pod="openshift-image-registry/image-registry-697d97f7c8-qw2fb" Dec 06 03:36:39 crc kubenswrapper[4980]: E1206 03:36:39.612311 4980 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 03:36:40.112300046 +0000 UTC m=+159.357051317 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qw2fb" (UID: "b25ba819-6e4e-4ca0-a570-c2726b8fbec5") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 03:36:39 crc kubenswrapper[4980]: I1206 03:36:39.626806 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-w6bmg" event={"ID":"dae23bd6-4c2b-441e-b44b-b6977bbd974b","Type":"ContainerStarted","Data":"9084c46659e61213107077fbd217502ebe8a0c17afa95763985b45e4b781fa28"} Dec 06 03:36:39 crc kubenswrapper[4980]: I1206 03:36:39.631067 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-4dcjh" event={"ID":"8be432d7-9e4b-4452-81df-91998326f92b","Type":"ContainerStarted","Data":"1ae726a8072285ed3c5927e954a65d906e9727055fa2cb572599314a492c6481"} Dec 06 03:36:39 crc kubenswrapper[4980]: I1206 03:36:39.637147 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-27gc7" event={"ID":"46290605-e114-4beb-b1c9-b0389f20f5f3","Type":"ContainerStarted","Data":"45e31ce432d890104b7023f191b1d2934096253fecac147e2e36649070658c79"} Dec 06 03:36:39 crc kubenswrapper[4980]: I1206 03:36:39.647808 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-qcn9f" event={"ID":"e1e4e0b4-0bfe-4b01-9c4e-710ba0ec03bd","Type":"ContainerStarted","Data":"31d907e5716469950c9151451fa388849f296c478c78c95d41a581c1835ef148"} Dec 06 03:36:39 crc kubenswrapper[4980]: I1206 03:36:39.647879 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-qcn9f" event={"ID":"e1e4e0b4-0bfe-4b01-9c4e-710ba0ec03bd","Type":"ContainerStarted","Data":"d76f9939f65dc7585caabe54d9478aaa3fc6771e52e8c2105dcf72fb1aa0845b"} Dec 06 03:36:39 crc kubenswrapper[4980]: I1206 03:36:39.703112 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-799j5" event={"ID":"3ee9a5b6-d84f-46d7-b3c2-aab85c634bd4","Type":"ContainerStarted","Data":"675786217dc50c49b33a3e7be6e59b8e6b35f07be65784738bd75b135360da75"} Dec 06 03:36:39 crc kubenswrapper[4980]: I1206 03:36:39.713051 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 03:36:39 crc kubenswrapper[4980]: E1206 03:36:39.714416 4980 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 03:36:40.214394044 +0000 UTC m=+159.459145315 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 03:36:39 crc kubenswrapper[4980]: I1206 03:36:39.721463 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29416530-9hm9m" event={"ID":"03704eb9-227d-4985-87cf-acd66ae76037","Type":"ContainerStarted","Data":"c4afa62f429f5f866be3d2100de061f502a4becc8ee55d85f991a5589586415c"} Dec 06 03:36:39 crc kubenswrapper[4980]: I1206 03:36:39.721577 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29416530-9hm9m" event={"ID":"03704eb9-227d-4985-87cf-acd66ae76037","Type":"ContainerStarted","Data":"bf422279b8c14c2fa296df47fdefd0d0a736c7eb49fc94d969d975ec797735c0"} Dec 06 03:36:39 crc kubenswrapper[4980]: I1206 03:36:39.754321 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-6h96g" event={"ID":"346bb011-c34e-4286-90d3-0242205e8e94","Type":"ContainerStarted","Data":"c1bf646e1a1a4bf331bae0dde027acdbea896ad1d5a0d6c940f5a5a5d3155843"} Dec 06 03:36:39 crc kubenswrapper[4980]: I1206 03:36:39.755043 4980 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/downloads-7954f5f757-6h96g" Dec 06 03:36:39 crc kubenswrapper[4980]: I1206 03:36:39.787337 4980 patch_prober.go:28] interesting pod/downloads-7954f5f757-6h96g container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.9:8080/\": dial tcp 10.217.0.9:8080: connect: connection refused" start-of-body= Dec 06 03:36:39 crc kubenswrapper[4980]: I1206 03:36:39.787647 4980 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-6h96g" podUID="346bb011-c34e-4286-90d3-0242205e8e94" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.9:8080/\": dial tcp 10.217.0.9:8080: connect: connection refused" Dec 06 03:36:39 crc kubenswrapper[4980]: I1206 03:36:39.787918 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-spd2v" event={"ID":"0a45bc54-be9c-4c4f-9cae-dc369d25be33","Type":"ContainerStarted","Data":"7a8808a8c7f810b0ac36ffaffe6515bcff3c06970f3d138268138750f476b32d"} Dec 06 03:36:39 crc kubenswrapper[4980]: I1206 03:36:39.815283 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qw2fb\" (UID: \"b25ba819-6e4e-4ca0-a570-c2726b8fbec5\") " pod="openshift-image-registry/image-registry-697d97f7c8-qw2fb" Dec 06 03:36:39 crc kubenswrapper[4980]: E1206 03:36:39.817126 4980 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 03:36:40.317110559 +0000 UTC m=+159.561861830 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qw2fb" (UID: "b25ba819-6e4e-4ca0-a570-c2726b8fbec5") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 03:36:39 crc kubenswrapper[4980]: I1206 03:36:39.817800 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-56s8p" event={"ID":"d7fb8e4e-88e7-4cc9-baf9-6e217813b28d","Type":"ContainerStarted","Data":"3bb65d364b0fa80f74f3921c2d7b39df0da360c8fc6c841bcb34f4b5ef0e524f"} Dec 06 03:36:39 crc kubenswrapper[4980]: I1206 03:36:39.817872 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-56s8p" event={"ID":"d7fb8e4e-88e7-4cc9-baf9-6e217813b28d","Type":"ContainerStarted","Data":"20c19db572f993b4403d3d5c2444904da4e375df907f343edccfd7b78cf1408e"} Dec 06 03:36:39 crc kubenswrapper[4980]: I1206 03:36:39.826449 4980 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/downloads-7954f5f757-6h96g" podStartSLOduration=137.826417158 podStartE2EDuration="2m17.826417158s" podCreationTimestamp="2025-12-06 03:34:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 03:36:39.825318486 +0000 UTC m=+159.070069757" watchObservedRunningTime="2025-12-06 03:36:39.826417158 +0000 UTC m=+159.071168429" Dec 06 03:36:39 crc kubenswrapper[4980]: I1206 03:36:39.826617 4980 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29416530-9hm9m" podStartSLOduration=137.826611423 podStartE2EDuration="2m17.826611423s" podCreationTimestamp="2025-12-06 03:34:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 03:36:39.751474394 +0000 UTC m=+158.996225665" watchObservedRunningTime="2025-12-06 03:36:39.826611423 +0000 UTC m=+159.071362694" Dec 06 03:36:39 crc kubenswrapper[4980]: I1206 03:36:39.845606 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-jp25g" event={"ID":"f2555ff9-efdb-449a-9f88-67a3c80ad0cf","Type":"ContainerStarted","Data":"17baae30a668f8fa6698e9149ae69701694e73a2f93b53198ea39f0724a988a3"} Dec 06 03:36:39 crc kubenswrapper[4980]: I1206 03:36:39.861085 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-cjvdn" event={"ID":"ccd9c9f8-3473-4f3d-884c-e545a0a502af","Type":"ContainerStarted","Data":"8c2fed3d71e4bc174ac60fbd32e83cf6439462bf76cd3aa7c37aa61eb360ab8d"} Dec 06 03:36:39 crc kubenswrapper[4980]: I1206 03:36:39.874277 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-r47r5" event={"ID":"b2da6efd-2b6d-4ca7-b056-32c4b132becf","Type":"ContainerStarted","Data":"c6ca7aed4ee93f12361376538dae48edc214ca514f64d40fe4f3f4314a4f7f7f"} Dec 06 03:36:39 crc kubenswrapper[4980]: I1206 03:36:39.906806 4980 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca-operator/service-ca-operator-777779d784-r47r5" podStartSLOduration=136.906786088 podStartE2EDuration="2m16.906786088s" podCreationTimestamp="2025-12-06 03:34:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 03:36:39.904571364 +0000 UTC m=+159.149322635" watchObservedRunningTime="2025-12-06 03:36:39.906786088 +0000 UTC m=+159.151537359" Dec 06 03:36:39 crc kubenswrapper[4980]: I1206 03:36:39.916820 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 03:36:39 crc kubenswrapper[4980]: E1206 03:36:39.917276 4980 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 03:36:40.41725454 +0000 UTC m=+159.662005811 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 03:36:39 crc kubenswrapper[4980]: I1206 03:36:39.917338 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qw2fb\" (UID: \"b25ba819-6e4e-4ca0-a570-c2726b8fbec5\") " pod="openshift-image-registry/image-registry-697d97f7c8-qw2fb" Dec 06 03:36:39 crc kubenswrapper[4980]: E1206 03:36:39.919376 4980 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 03:36:40.41934568 +0000 UTC m=+159.664096951 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qw2fb" (UID: "b25ba819-6e4e-4ca0-a570-c2726b8fbec5") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 03:36:40 crc kubenswrapper[4980]: I1206 03:36:39.994625 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-84ft2" event={"ID":"a3673c8a-c303-42ac-8565-f03e5d457aa1","Type":"ContainerStarted","Data":"dd24a518df7d382ef6c0f54acd0ba7840ac7a6c8063d1a669944d6f297c2875d"} Dec 06 03:36:40 crc kubenswrapper[4980]: I1206 03:36:40.006747 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-8jc8f" event={"ID":"c7146eb0-acc5-4bac-b429-e1e00761e666","Type":"ContainerStarted","Data":"b4b5048b9b84ba93932f563231667fc163ba0556d51a30a00a674ed654e23ceb"} Dec 06 03:36:40 crc kubenswrapper[4980]: I1206 03:36:40.025279 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 03:36:40 crc kubenswrapper[4980]: E1206 03:36:40.025372 4980 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 03:36:40.525358881 +0000 UTC m=+159.770110152 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 03:36:40 crc kubenswrapper[4980]: I1206 03:36:40.025754 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-qpf6b" event={"ID":"2692cb89-853a-4c4e-a6a6-570673db6f14","Type":"ContainerStarted","Data":"d004a4a26392f0b09e3467e76cd873eb79abed10e92bbae92ae704997a565a4a"} Dec 06 03:36:40 crc kubenswrapper[4980]: I1206 03:36:40.025877 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qw2fb\" (UID: \"b25ba819-6e4e-4ca0-a570-c2726b8fbec5\") " pod="openshift-image-registry/image-registry-697d97f7c8-qw2fb" Dec 06 03:36:40 crc kubenswrapper[4980]: E1206 03:36:40.028432 4980 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 03:36:40.528422009 +0000 UTC m=+159.773173280 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qw2fb" (UID: "b25ba819-6e4e-4ca0-a570-c2726b8fbec5") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 03:36:40 crc kubenswrapper[4980]: I1206 03:36:40.034924 4980 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication-operator/authentication-operator-69f744f599-84ft2" podStartSLOduration=138.034906487 podStartE2EDuration="2m18.034906487s" podCreationTimestamp="2025-12-06 03:34:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 03:36:40.032551349 +0000 UTC m=+159.277302620" watchObservedRunningTime="2025-12-06 03:36:40.034906487 +0000 UTC m=+159.279657758" Dec 06 03:36:40 crc kubenswrapper[4980]: I1206 03:36:40.053686 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-49749" event={"ID":"05aaf766-565d-4b05-97ec-6c8078d33f85","Type":"ContainerStarted","Data":"af46dc42986ae374a950452b24939ee06c6f2be1e2c67f66312f0a95adb3ae36"} Dec 06 03:36:40 crc kubenswrapper[4980]: I1206 03:36:40.063801 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-74hx2" event={"ID":"9b9bddd5-f192-4b4c-81ee-b523bf9c8705","Type":"ContainerStarted","Data":"c969a8f2fefbc4c8d68337de5c159d66ff233d1b2cbd1b3624dd8a0c044d9183"} Dec 06 03:36:40 crc kubenswrapper[4980]: I1206 03:36:40.074491 4980 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-8jc8f" podStartSLOduration=137.074474259 podStartE2EDuration="2m17.074474259s" podCreationTimestamp="2025-12-06 03:34:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 03:36:40.072972966 +0000 UTC m=+159.317724237" watchObservedRunningTime="2025-12-06 03:36:40.074474259 +0000 UTC m=+159.319225530" Dec 06 03:36:40 crc kubenswrapper[4980]: I1206 03:36:40.097762 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-cqk6j" event={"ID":"e81d9a6d-991f-46b0-aba5-6f76a5b1031a","Type":"ContainerStarted","Data":"ef438880ccaed95b2f07e4f730222adab3a4e9e71185427ffd2a257698a93b35"} Dec 06 03:36:40 crc kubenswrapper[4980]: I1206 03:36:40.111707 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-gjdsq" event={"ID":"949f499a-5680-4fec-ab59-5ced4f24ba2f","Type":"ContainerStarted","Data":"34a26ff32bf2986cf923b8e76969adf9fd4c2f9a6318fffbc2d49a498d3d29a9"} Dec 06 03:36:40 crc kubenswrapper[4980]: I1206 03:36:40.121082 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-tlpd8" event={"ID":"d1d006b0-92b0-45d5-9c28-fd43f879b310","Type":"ContainerStarted","Data":"b8dbf0314ab9ae97b1b4db31f4aeaa3585e8f4333056ca1ebbffe80f2a93770d"} Dec 06 03:36:40 crc kubenswrapper[4980]: I1206 03:36:40.124355 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-hjrkf" event={"ID":"a7ee597a-a4fc-45ef-841f-1d4f666ccfcf","Type":"ContainerStarted","Data":"9e58f6ce1c230befe779f78c68b743593fa4684bfa6e974d7b099155a55746ad"} Dec 06 03:36:40 crc kubenswrapper[4980]: I1206 03:36:40.125375 4980 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-hjrkf" Dec 06 03:36:40 crc kubenswrapper[4980]: I1206 03:36:40.130691 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 03:36:40 crc kubenswrapper[4980]: E1206 03:36:40.130819 4980 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 03:36:40.630790905 +0000 UTC m=+159.875542176 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 03:36:40 crc kubenswrapper[4980]: I1206 03:36:40.131091 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qw2fb\" (UID: \"b25ba819-6e4e-4ca0-a570-c2726b8fbec5\") " pod="openshift-image-registry/image-registry-697d97f7c8-qw2fb" Dec 06 03:36:40 crc kubenswrapper[4980]: E1206 03:36:40.131525 4980 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 03:36:40.631497305 +0000 UTC m=+159.876248576 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qw2fb" (UID: "b25ba819-6e4e-4ca0-a570-c2726b8fbec5") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 03:36:40 crc kubenswrapper[4980]: I1206 03:36:40.169427 4980 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-cqk6j" podStartSLOduration=137.16940342 podStartE2EDuration="2m17.16940342s" podCreationTimestamp="2025-12-06 03:34:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 03:36:40.141163684 +0000 UTC m=+159.385914955" watchObservedRunningTime="2025-12-06 03:36:40.16940342 +0000 UTC m=+159.414154691" Dec 06 03:36:40 crc kubenswrapper[4980]: I1206 03:36:40.193661 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-xjtnv" event={"ID":"f08eff3e-89fe-4caf-9ec1-b32ee99f3c10","Type":"ContainerStarted","Data":"bba192cecfb3403009159b70fc76095bed48b1f0d25e884084cf448ee09cee33"} Dec 06 03:36:40 crc kubenswrapper[4980]: I1206 03:36:40.194696 4980 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-hjrkf" podStartSLOduration=137.194659209 podStartE2EDuration="2m17.194659209s" podCreationTimestamp="2025-12-06 03:34:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 03:36:40.193383102 +0000 UTC m=+159.438134373" watchObservedRunningTime="2025-12-06 03:36:40.194659209 +0000 UTC m=+159.439410480" Dec 06 03:36:40 crc kubenswrapper[4980]: I1206 03:36:40.221612 4980 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-xjtnv" podStartSLOduration=137.221592536 podStartE2EDuration="2m17.221592536s" podCreationTimestamp="2025-12-06 03:34:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 03:36:40.221462102 +0000 UTC m=+159.466213373" watchObservedRunningTime="2025-12-06 03:36:40.221592536 +0000 UTC m=+159.466343807" Dec 06 03:36:40 crc kubenswrapper[4980]: I1206 03:36:40.236261 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 03:36:40 crc kubenswrapper[4980]: E1206 03:36:40.238202 4980 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 03:36:40.737531126 +0000 UTC m=+159.982282397 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 03:36:40 crc kubenswrapper[4980]: I1206 03:36:40.240779 4980 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-hjrkf" Dec 06 03:36:40 crc kubenswrapper[4980]: I1206 03:36:40.383900 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qw2fb\" (UID: \"b25ba819-6e4e-4ca0-a570-c2726b8fbec5\") " pod="openshift-image-registry/image-registry-697d97f7c8-qw2fb" Dec 06 03:36:40 crc kubenswrapper[4980]: E1206 03:36:40.385970 4980 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 03:36:40.88595761 +0000 UTC m=+160.130708881 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qw2fb" (UID: "b25ba819-6e4e-4ca0-a570-c2726b8fbec5") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 03:36:40 crc kubenswrapper[4980]: I1206 03:36:40.436930 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-d6dd4" event={"ID":"abfe3c42-9229-486f-8c87-8b883f6821e0","Type":"ContainerStarted","Data":"2371aafce6d068cb94b0326470287184ae6746fcc7872002bb280443e1381a98"} Dec 06 03:36:40 crc kubenswrapper[4980]: I1206 03:36:40.519915 4980 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-558db77b4-pmmcw" Dec 06 03:36:40 crc kubenswrapper[4980]: I1206 03:36:40.520626 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 03:36:40 crc kubenswrapper[4980]: E1206 03:36:40.520743 4980 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 03:36:41.020725425 +0000 UTC m=+160.265476696 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 03:36:40 crc kubenswrapper[4980]: I1206 03:36:40.521154 4980 patch_prober.go:28] interesting pod/router-default-5444994796-pzzkd container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 06 03:36:40 crc kubenswrapper[4980]: [-]has-synced failed: reason withheld Dec 06 03:36:40 crc kubenswrapper[4980]: [+]process-running ok Dec 06 03:36:40 crc kubenswrapper[4980]: healthz check failed Dec 06 03:36:40 crc kubenswrapper[4980]: I1206 03:36:40.521200 4980 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-pzzkd" podUID="acb53ac6-3f09-424c-a830-a1e15eec0d62" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 06 03:36:40 crc kubenswrapper[4980]: I1206 03:36:40.521493 4980 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console-operator/console-operator-58897d9998-kftgf" Dec 06 03:36:40 crc kubenswrapper[4980]: I1206 03:36:40.521812 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qw2fb\" (UID: \"b25ba819-6e4e-4ca0-a570-c2726b8fbec5\") " pod="openshift-image-registry/image-registry-697d97f7c8-qw2fb" Dec 06 03:36:40 crc kubenswrapper[4980]: E1206 03:36:40.524920 4980 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 03:36:41.024902985 +0000 UTC m=+160.269654246 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qw2fb" (UID: "b25ba819-6e4e-4ca0-a570-c2726b8fbec5") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 03:36:40 crc kubenswrapper[4980]: I1206 03:36:40.650688 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 03:36:40 crc kubenswrapper[4980]: E1206 03:36:40.651239 4980 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 03:36:41.151211779 +0000 UTC m=+160.395963050 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 03:36:40 crc kubenswrapper[4980]: I1206 03:36:40.761592 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qw2fb\" (UID: \"b25ba819-6e4e-4ca0-a570-c2726b8fbec5\") " pod="openshift-image-registry/image-registry-697d97f7c8-qw2fb" Dec 06 03:36:40 crc kubenswrapper[4980]: E1206 03:36:40.762033 4980 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 03:36:41.262014468 +0000 UTC m=+160.506765739 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qw2fb" (UID: "b25ba819-6e4e-4ca0-a570-c2726b8fbec5") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 03:36:40 crc kubenswrapper[4980]: I1206 03:36:40.862489 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 03:36:40 crc kubenswrapper[4980]: E1206 03:36:40.863168 4980 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 03:36:41.363152718 +0000 UTC m=+160.607903989 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 03:36:40 crc kubenswrapper[4980]: I1206 03:36:40.964491 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qw2fb\" (UID: \"b25ba819-6e4e-4ca0-a570-c2726b8fbec5\") " pod="openshift-image-registry/image-registry-697d97f7c8-qw2fb" Dec 06 03:36:40 crc kubenswrapper[4980]: E1206 03:36:40.965369 4980 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 03:36:41.465355658 +0000 UTC m=+160.710106929 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qw2fb" (UID: "b25ba819-6e4e-4ca0-a570-c2726b8fbec5") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 03:36:41 crc kubenswrapper[4980]: I1206 03:36:41.066396 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 03:36:41 crc kubenswrapper[4980]: E1206 03:36:41.066615 4980 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 03:36:41.566599871 +0000 UTC m=+160.811351132 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 03:36:41 crc kubenswrapper[4980]: I1206 03:36:41.167377 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qw2fb\" (UID: \"b25ba819-6e4e-4ca0-a570-c2726b8fbec5\") " pod="openshift-image-registry/image-registry-697d97f7c8-qw2fb" Dec 06 03:36:41 crc kubenswrapper[4980]: E1206 03:36:41.167801 4980 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 03:36:41.667786702 +0000 UTC m=+160.912537973 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qw2fb" (UID: "b25ba819-6e4e-4ca0-a570-c2726b8fbec5") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 03:36:41 crc kubenswrapper[4980]: I1206 03:36:41.270414 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 03:36:41 crc kubenswrapper[4980]: E1206 03:36:41.270741 4980 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 03:36:41.770722314 +0000 UTC m=+161.015473575 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 03:36:41 crc kubenswrapper[4980]: I1206 03:36:41.375302 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qw2fb\" (UID: \"b25ba819-6e4e-4ca0-a570-c2726b8fbec5\") " pod="openshift-image-registry/image-registry-697d97f7c8-qw2fb" Dec 06 03:36:41 crc kubenswrapper[4980]: E1206 03:36:41.375749 4980 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 03:36:41.875734776 +0000 UTC m=+161.120486057 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qw2fb" (UID: "b25ba819-6e4e-4ca0-a570-c2726b8fbec5") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 03:36:41 crc kubenswrapper[4980]: I1206 03:36:41.440925 4980 patch_prober.go:28] interesting pod/router-default-5444994796-pzzkd container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 06 03:36:41 crc kubenswrapper[4980]: [-]has-synced failed: reason withheld Dec 06 03:36:41 crc kubenswrapper[4980]: [+]process-running ok Dec 06 03:36:41 crc kubenswrapper[4980]: healthz check failed Dec 06 03:36:41 crc kubenswrapper[4980]: I1206 03:36:41.440980 4980 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-pzzkd" podUID="acb53ac6-3f09-424c-a830-a1e15eec0d62" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 06 03:36:41 crc kubenswrapper[4980]: I1206 03:36:41.454361 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-cjvdn" event={"ID":"ccd9c9f8-3473-4f3d-884c-e545a0a502af","Type":"ContainerStarted","Data":"377caca5dbfb6d37bcbd8c5f91671ea6f50f001cf156679209e3e1b618d3d81b"} Dec 06 03:36:41 crc kubenswrapper[4980]: I1206 03:36:41.468103 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-sq6bp" event={"ID":"9476a6a0-4d40-4c85-9bd1-2162c971e3dd","Type":"ContainerStarted","Data":"2922be787f767a73db35f5788b3a542756c6392335e46ed62fa7c55044d0d7cb"} Dec 06 03:36:41 crc kubenswrapper[4980]: I1206 03:36:41.469133 4980 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-sq6bp" Dec 06 03:36:41 crc kubenswrapper[4980]: I1206 03:36:41.474472 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-56s8p" event={"ID":"d7fb8e4e-88e7-4cc9-baf9-6e217813b28d","Type":"ContainerStarted","Data":"df4670a01bbd2af3907318509ee7b1409da2a144cb26fce7b067390e019cb7e1"} Dec 06 03:36:41 crc kubenswrapper[4980]: I1206 03:36:41.476127 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 03:36:41 crc kubenswrapper[4980]: E1206 03:36:41.477047 4980 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 03:36:41.97703165 +0000 UTC m=+161.221782921 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 03:36:41 crc kubenswrapper[4980]: I1206 03:36:41.479595 4980 patch_prober.go:28] interesting pod/olm-operator-6b444d44fb-sq6bp container/olm-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.26:8443/healthz\": dial tcp 10.217.0.26:8443: connect: connection refused" start-of-body= Dec 06 03:36:41 crc kubenswrapper[4980]: I1206 03:36:41.479719 4980 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-sq6bp" podUID="9476a6a0-4d40-4c85-9bd1-2162c971e3dd" containerName="olm-operator" probeResult="failure" output="Get \"https://10.217.0.26:8443/healthz\": dial tcp 10.217.0.26:8443: connect: connection refused" Dec 06 03:36:41 crc kubenswrapper[4980]: I1206 03:36:41.494734 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-tlpd8" event={"ID":"d1d006b0-92b0-45d5-9c28-fd43f879b310","Type":"ContainerStarted","Data":"a6adff7393183a97388856b5914d6eb15d63c08326e19a68e064d4df159b8ab9"} Dec 06 03:36:41 crc kubenswrapper[4980]: I1206 03:36:41.494781 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-tlpd8" event={"ID":"d1d006b0-92b0-45d5-9c28-fd43f879b310","Type":"ContainerStarted","Data":"b136ed88d88143fa8e3a22e07d2baae8a51ffee53900cd49725ac814db33122a"} Dec 06 03:36:41 crc kubenswrapper[4980]: I1206 03:36:41.497774 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-84ft2" event={"ID":"a3673c8a-c303-42ac-8565-f03e5d457aa1","Type":"ContainerStarted","Data":"5b02fb4c4ef5446a0a0c625b075b2aed67835bf36f2aa03ed09f67555f535e4a"} Dec 06 03:36:41 crc kubenswrapper[4980]: I1206 03:36:41.501982 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-74hx2" event={"ID":"9b9bddd5-f192-4b4c-81ee-b523bf9c8705","Type":"ContainerStarted","Data":"450d8ca7fc6f0d8becaecbf41796718832f6c18c0ce6113503423958021beeed"} Dec 06 03:36:41 crc kubenswrapper[4980]: I1206 03:36:41.544085 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-4dcjh" event={"ID":"8be432d7-9e4b-4452-81df-91998326f92b","Type":"ContainerStarted","Data":"12243bf8dc3e434a6a117c55c58d9ba8f9fd7604e65af96e9bf26504a5e7eb03"} Dec 06 03:36:41 crc kubenswrapper[4980]: I1206 03:36:41.547553 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-jp25g" event={"ID":"f2555ff9-efdb-449a-9f88-67a3c80ad0cf","Type":"ContainerStarted","Data":"ee27c77a52c23be64c82847feaa9853b86e304a3d49231e21e924fbea8c32d07"} Dec 06 03:36:41 crc kubenswrapper[4980]: I1206 03:36:41.552856 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-d6dd4" event={"ID":"abfe3c42-9229-486f-8c87-8b883f6821e0","Type":"ContainerStarted","Data":"09d2385eafa22cfe1dcf1225665993b3da91c40d834e993f92ed144710edc635"} Dec 06 03:36:41 crc kubenswrapper[4980]: I1206 03:36:41.552900 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-d6dd4" event={"ID":"abfe3c42-9229-486f-8c87-8b883f6821e0","Type":"ContainerStarted","Data":"41a643ab29bf28dbadfd0e49f9f60968f2a2356247bb6001b3067cffd4ccdd1f"} Dec 06 03:36:41 crc kubenswrapper[4980]: I1206 03:36:41.554627 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-r47r5" event={"ID":"b2da6efd-2b6d-4ca7-b056-32c4b132becf","Type":"ContainerStarted","Data":"cf2ee35474e21d81c5df1ddd761cf7143928a097257c838e85ecc2a1c5710682"} Dec 06 03:36:41 crc kubenswrapper[4980]: I1206 03:36:41.556556 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-w6bmg" event={"ID":"dae23bd6-4c2b-441e-b44b-b6977bbd974b","Type":"ContainerStarted","Data":"4f188ffbc7154ef76ea38d68af97f8aeb2d013a29dbcc98f523e94dcfef58406"} Dec 06 03:36:41 crc kubenswrapper[4980]: I1206 03:36:41.558398 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-s6h9g" event={"ID":"2cf31abb-ef18-4ba5-8cc9-bf3f7507f771","Type":"ContainerStarted","Data":"603febf3d78aa47efe6f3232f4495a7312a6e3b5dd740656ad577ed32fec6f36"} Dec 06 03:36:41 crc kubenswrapper[4980]: I1206 03:36:41.560617 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-qpf6b" event={"ID":"2692cb89-853a-4c4e-a6a6-570673db6f14","Type":"ContainerStarted","Data":"50d031163bb91e828bfec7fdbef8c96d2b00e5f6c73705334048fbcbf40f353b"} Dec 06 03:36:41 crc kubenswrapper[4980]: I1206 03:36:41.562449 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-49749" event={"ID":"05aaf766-565d-4b05-97ec-6c8078d33f85","Type":"ContainerStarted","Data":"2699acb0b787c94949fbeb511ca04947d8e8f6441fbb8eaf4645eb5939cb4751"} Dec 06 03:36:41 crc kubenswrapper[4980]: I1206 03:36:41.562476 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-49749" event={"ID":"05aaf766-565d-4b05-97ec-6c8078d33f85","Type":"ContainerStarted","Data":"2ab2b482879710618399579e9a055b296919222eb2c131a571aff49d485b035e"} Dec 06 03:36:41 crc kubenswrapper[4980]: I1206 03:36:41.569469 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-qcn9f" event={"ID":"e1e4e0b4-0bfe-4b01-9c4e-710ba0ec03bd","Type":"ContainerStarted","Data":"20e968e0db39907dd07b8224ae813a89b8cd62226dc0db0479296ef8be7bf342"} Dec 06 03:36:41 crc kubenswrapper[4980]: I1206 03:36:41.569961 4980 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-qcn9f" Dec 06 03:36:41 crc kubenswrapper[4980]: I1206 03:36:41.571389 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-gjdsq" event={"ID":"949f499a-5680-4fec-ab59-5ced4f24ba2f","Type":"ContainerStarted","Data":"bd7806d9430d8cb5779e92902993d13757d80cd7f66b06eedb1033832fd9eef2"} Dec 06 03:36:41 crc kubenswrapper[4980]: I1206 03:36:41.571899 4980 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-gjdsq" Dec 06 03:36:41 crc kubenswrapper[4980]: I1206 03:36:41.577201 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qw2fb\" (UID: \"b25ba819-6e4e-4ca0-a570-c2726b8fbec5\") " pod="openshift-image-registry/image-registry-697d97f7c8-qw2fb" Dec 06 03:36:41 crc kubenswrapper[4980]: E1206 03:36:41.579406 4980 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 03:36:42.079395366 +0000 UTC m=+161.324146637 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qw2fb" (UID: "b25ba819-6e4e-4ca0-a570-c2726b8fbec5") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 03:36:41 crc kubenswrapper[4980]: I1206 03:36:41.579870 4980 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-gjdsq container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.43:5443/healthz\": dial tcp 10.217.0.43:5443: connect: connection refused" start-of-body= Dec 06 03:36:41 crc kubenswrapper[4980]: I1206 03:36:41.579952 4980 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-gjdsq" podUID="949f499a-5680-4fec-ab59-5ced4f24ba2f" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.43:5443/healthz\": dial tcp 10.217.0.43:5443: connect: connection refused" Dec 06 03:36:41 crc kubenswrapper[4980]: I1206 03:36:41.585896 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-spd2v" event={"ID":"0a45bc54-be9c-4c4f-9cae-dc369d25be33","Type":"ContainerStarted","Data":"b6c5027c91c80a97467c1e8c50318fab43b7ec468acda83d6d1f8288d5c58543"} Dec 06 03:36:41 crc kubenswrapper[4980]: I1206 03:36:41.586759 4980 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-spd2v" Dec 06 03:36:41 crc kubenswrapper[4980]: I1206 03:36:41.589118 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-27gc7" event={"ID":"46290605-e114-4beb-b1c9-b0389f20f5f3","Type":"ContainerStarted","Data":"146eb98cb9e50da754562964eda637fbee4838570648c2e1fe0233d8b38ec633"} Dec 06 03:36:41 crc kubenswrapper[4980]: I1206 03:36:41.589219 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-27gc7" event={"ID":"46290605-e114-4beb-b1c9-b0389f20f5f3","Type":"ContainerStarted","Data":"d42ba999c93381e5c7964b2c1a619513e0adad11daaa52f30c7a3f25807fdda2"} Dec 06 03:36:41 crc kubenswrapper[4980]: I1206 03:36:41.596754 4980 patch_prober.go:28] interesting pod/downloads-7954f5f757-6h96g container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.9:8080/\": dial tcp 10.217.0.9:8080: connect: connection refused" start-of-body= Dec 06 03:36:41 crc kubenswrapper[4980]: I1206 03:36:41.597099 4980 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-6h96g" podUID="346bb011-c34e-4286-90d3-0242205e8e94" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.9:8080/\": dial tcp 10.217.0.9:8080: connect: connection refused" Dec 06 03:36:41 crc kubenswrapper[4980]: I1206 03:36:41.597427 4980 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-spd2v container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.23:8080/healthz\": dial tcp 10.217.0.23:8080: connect: connection refused" start-of-body= Dec 06 03:36:41 crc kubenswrapper[4980]: I1206 03:36:41.597688 4980 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-spd2v" podUID="0a45bc54-be9c-4c4f-9cae-dc369d25be33" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.23:8080/healthz\": dial tcp 10.217.0.23:8080: connect: connection refused" Dec 06 03:36:41 crc kubenswrapper[4980]: I1206 03:36:41.679432 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 03:36:41 crc kubenswrapper[4980]: E1206 03:36:41.681044 4980 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 03:36:42.181025759 +0000 UTC m=+161.425777030 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 03:36:41 crc kubenswrapper[4980]: I1206 03:36:41.834483 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qw2fb\" (UID: \"b25ba819-6e4e-4ca0-a570-c2726b8fbec5\") " pod="openshift-image-registry/image-registry-697d97f7c8-qw2fb" Dec 06 03:36:41 crc kubenswrapper[4980]: E1206 03:36:41.835479 4980 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 03:36:42.335461868 +0000 UTC m=+161.580213139 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qw2fb" (UID: "b25ba819-6e4e-4ca0-a570-c2726b8fbec5") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 03:36:41 crc kubenswrapper[4980]: I1206 03:36:41.935431 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 03:36:41 crc kubenswrapper[4980]: E1206 03:36:41.935826 4980 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 03:36:42.435810615 +0000 UTC m=+161.680561886 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 03:36:41 crc kubenswrapper[4980]: I1206 03:36:41.944905 4980 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-mkd2m"] Dec 06 03:36:41 crc kubenswrapper[4980]: I1206 03:36:41.946036 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-mkd2m" Dec 06 03:36:41 crc kubenswrapper[4980]: I1206 03:36:41.959801 4980 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Dec 06 03:36:41 crc kubenswrapper[4980]: I1206 03:36:41.960615 4980 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-58d94"] Dec 06 03:36:41 crc kubenswrapper[4980]: I1206 03:36:41.961740 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-58d94" Dec 06 03:36:41 crc kubenswrapper[4980]: I1206 03:36:41.976134 4980 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-mkd2m"] Dec 06 03:36:41 crc kubenswrapper[4980]: I1206 03:36:41.981999 4980 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Dec 06 03:36:41 crc kubenswrapper[4980]: I1206 03:36:41.992080 4980 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-58d94"] Dec 06 03:36:42 crc kubenswrapper[4980]: I1206 03:36:42.037276 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9c8b3374-77eb-44ad-b270-2263c6c23e28-utilities\") pod \"community-operators-mkd2m\" (UID: \"9c8b3374-77eb-44ad-b270-2263c6c23e28\") " pod="openshift-marketplace/community-operators-mkd2m" Dec 06 03:36:42 crc kubenswrapper[4980]: I1206 03:36:42.037319 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ppcll\" (UniqueName: \"kubernetes.io/projected/9c8b3374-77eb-44ad-b270-2263c6c23e28-kube-api-access-ppcll\") pod \"community-operators-mkd2m\" (UID: \"9c8b3374-77eb-44ad-b270-2263c6c23e28\") " pod="openshift-marketplace/community-operators-mkd2m" Dec 06 03:36:42 crc kubenswrapper[4980]: I1206 03:36:42.037370 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9c8b3374-77eb-44ad-b270-2263c6c23e28-catalog-content\") pod \"community-operators-mkd2m\" (UID: \"9c8b3374-77eb-44ad-b270-2263c6c23e28\") " pod="openshift-marketplace/community-operators-mkd2m" Dec 06 03:36:42 crc kubenswrapper[4980]: I1206 03:36:42.037393 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qw2fb\" (UID: \"b25ba819-6e4e-4ca0-a570-c2726b8fbec5\") " pod="openshift-image-registry/image-registry-697d97f7c8-qw2fb" Dec 06 03:36:42 crc kubenswrapper[4980]: I1206 03:36:42.037420 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e965bb7a-1bf0-4574-aae4-2a45c75fb11c-catalog-content\") pod \"certified-operators-58d94\" (UID: \"e965bb7a-1bf0-4574-aae4-2a45c75fb11c\") " pod="openshift-marketplace/certified-operators-58d94" Dec 06 03:36:42 crc kubenswrapper[4980]: I1206 03:36:42.037440 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e965bb7a-1bf0-4574-aae4-2a45c75fb11c-utilities\") pod \"certified-operators-58d94\" (UID: \"e965bb7a-1bf0-4574-aae4-2a45c75fb11c\") " pod="openshift-marketplace/certified-operators-58d94" Dec 06 03:36:42 crc kubenswrapper[4980]: I1206 03:36:42.037457 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4qvrf\" (UniqueName: \"kubernetes.io/projected/e965bb7a-1bf0-4574-aae4-2a45c75fb11c-kube-api-access-4qvrf\") pod \"certified-operators-58d94\" (UID: \"e965bb7a-1bf0-4574-aae4-2a45c75fb11c\") " pod="openshift-marketplace/certified-operators-58d94" Dec 06 03:36:42 crc kubenswrapper[4980]: E1206 03:36:42.037841 4980 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 03:36:42.53782956 +0000 UTC m=+161.782580831 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qw2fb" (UID: "b25ba819-6e4e-4ca0-a570-c2726b8fbec5") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 03:36:42 crc kubenswrapper[4980]: I1206 03:36:42.048227 4980 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-2vfzc"] Dec 06 03:36:42 crc kubenswrapper[4980]: I1206 03:36:42.049302 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-2vfzc" Dec 06 03:36:42 crc kubenswrapper[4980]: I1206 03:36:42.104964 4980 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-2vfzc"] Dec 06 03:36:42 crc kubenswrapper[4980]: I1206 03:36:42.138448 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 03:36:42 crc kubenswrapper[4980]: I1206 03:36:42.138855 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e965bb7a-1bf0-4574-aae4-2a45c75fb11c-catalog-content\") pod \"certified-operators-58d94\" (UID: \"e965bb7a-1bf0-4574-aae4-2a45c75fb11c\") " pod="openshift-marketplace/certified-operators-58d94" Dec 06 03:36:42 crc kubenswrapper[4980]: I1206 03:36:42.138898 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e965bb7a-1bf0-4574-aae4-2a45c75fb11c-utilities\") pod \"certified-operators-58d94\" (UID: \"e965bb7a-1bf0-4574-aae4-2a45c75fb11c\") " pod="openshift-marketplace/certified-operators-58d94" Dec 06 03:36:42 crc kubenswrapper[4980]: I1206 03:36:42.138958 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4qvrf\" (UniqueName: \"kubernetes.io/projected/e965bb7a-1bf0-4574-aae4-2a45c75fb11c-kube-api-access-4qvrf\") pod \"certified-operators-58d94\" (UID: \"e965bb7a-1bf0-4574-aae4-2a45c75fb11c\") " pod="openshift-marketplace/certified-operators-58d94" Dec 06 03:36:42 crc kubenswrapper[4980]: I1206 03:36:42.139014 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wxr7g\" (UniqueName: \"kubernetes.io/projected/652a7feb-30e3-4e9f-b1a9-c8ef201c658c-kube-api-access-wxr7g\") pod \"community-operators-2vfzc\" (UID: \"652a7feb-30e3-4e9f-b1a9-c8ef201c658c\") " pod="openshift-marketplace/community-operators-2vfzc" Dec 06 03:36:42 crc kubenswrapper[4980]: I1206 03:36:42.139055 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9c8b3374-77eb-44ad-b270-2263c6c23e28-utilities\") pod \"community-operators-mkd2m\" (UID: \"9c8b3374-77eb-44ad-b270-2263c6c23e28\") " pod="openshift-marketplace/community-operators-mkd2m" Dec 06 03:36:42 crc kubenswrapper[4980]: I1206 03:36:42.139079 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ppcll\" (UniqueName: \"kubernetes.io/projected/9c8b3374-77eb-44ad-b270-2263c6c23e28-kube-api-access-ppcll\") pod \"community-operators-mkd2m\" (UID: \"9c8b3374-77eb-44ad-b270-2263c6c23e28\") " pod="openshift-marketplace/community-operators-mkd2m" Dec 06 03:36:42 crc kubenswrapper[4980]: I1206 03:36:42.139119 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/652a7feb-30e3-4e9f-b1a9-c8ef201c658c-catalog-content\") pod \"community-operators-2vfzc\" (UID: \"652a7feb-30e3-4e9f-b1a9-c8ef201c658c\") " pod="openshift-marketplace/community-operators-2vfzc" Dec 06 03:36:42 crc kubenswrapper[4980]: I1206 03:36:42.139163 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/652a7feb-30e3-4e9f-b1a9-c8ef201c658c-utilities\") pod \"community-operators-2vfzc\" (UID: \"652a7feb-30e3-4e9f-b1a9-c8ef201c658c\") " pod="openshift-marketplace/community-operators-2vfzc" Dec 06 03:36:42 crc kubenswrapper[4980]: I1206 03:36:42.139189 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9c8b3374-77eb-44ad-b270-2263c6c23e28-catalog-content\") pod \"community-operators-mkd2m\" (UID: \"9c8b3374-77eb-44ad-b270-2263c6c23e28\") " pod="openshift-marketplace/community-operators-mkd2m" Dec 06 03:36:42 crc kubenswrapper[4980]: I1206 03:36:42.139762 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9c8b3374-77eb-44ad-b270-2263c6c23e28-catalog-content\") pod \"community-operators-mkd2m\" (UID: \"9c8b3374-77eb-44ad-b270-2263c6c23e28\") " pod="openshift-marketplace/community-operators-mkd2m" Dec 06 03:36:42 crc kubenswrapper[4980]: E1206 03:36:42.139871 4980 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 03:36:42.639841075 +0000 UTC m=+161.884592346 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 03:36:42 crc kubenswrapper[4980]: I1206 03:36:42.140193 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e965bb7a-1bf0-4574-aae4-2a45c75fb11c-catalog-content\") pod \"certified-operators-58d94\" (UID: \"e965bb7a-1bf0-4574-aae4-2a45c75fb11c\") " pod="openshift-marketplace/certified-operators-58d94" Dec 06 03:36:42 crc kubenswrapper[4980]: I1206 03:36:42.140729 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e965bb7a-1bf0-4574-aae4-2a45c75fb11c-utilities\") pod \"certified-operators-58d94\" (UID: \"e965bb7a-1bf0-4574-aae4-2a45c75fb11c\") " pod="openshift-marketplace/certified-operators-58d94" Dec 06 03:36:42 crc kubenswrapper[4980]: I1206 03:36:42.141369 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9c8b3374-77eb-44ad-b270-2263c6c23e28-utilities\") pod \"community-operators-mkd2m\" (UID: \"9c8b3374-77eb-44ad-b270-2263c6c23e28\") " pod="openshift-marketplace/community-operators-mkd2m" Dec 06 03:36:42 crc kubenswrapper[4980]: I1206 03:36:42.190457 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ppcll\" (UniqueName: \"kubernetes.io/projected/9c8b3374-77eb-44ad-b270-2263c6c23e28-kube-api-access-ppcll\") pod \"community-operators-mkd2m\" (UID: \"9c8b3374-77eb-44ad-b270-2263c6c23e28\") " pod="openshift-marketplace/community-operators-mkd2m" Dec 06 03:36:42 crc kubenswrapper[4980]: I1206 03:36:42.215572 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4qvrf\" (UniqueName: \"kubernetes.io/projected/e965bb7a-1bf0-4574-aae4-2a45c75fb11c-kube-api-access-4qvrf\") pod \"certified-operators-58d94\" (UID: \"e965bb7a-1bf0-4574-aae4-2a45c75fb11c\") " pod="openshift-marketplace/certified-operators-58d94" Dec 06 03:36:42 crc kubenswrapper[4980]: I1206 03:36:42.242124 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wxr7g\" (UniqueName: \"kubernetes.io/projected/652a7feb-30e3-4e9f-b1a9-c8ef201c658c-kube-api-access-wxr7g\") pod \"community-operators-2vfzc\" (UID: \"652a7feb-30e3-4e9f-b1a9-c8ef201c658c\") " pod="openshift-marketplace/community-operators-2vfzc" Dec 06 03:36:42 crc kubenswrapper[4980]: I1206 03:36:42.242396 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/652a7feb-30e3-4e9f-b1a9-c8ef201c658c-catalog-content\") pod \"community-operators-2vfzc\" (UID: \"652a7feb-30e3-4e9f-b1a9-c8ef201c658c\") " pod="openshift-marketplace/community-operators-2vfzc" Dec 06 03:36:42 crc kubenswrapper[4980]: I1206 03:36:42.242483 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/652a7feb-30e3-4e9f-b1a9-c8ef201c658c-utilities\") pod \"community-operators-2vfzc\" (UID: \"652a7feb-30e3-4e9f-b1a9-c8ef201c658c\") " pod="openshift-marketplace/community-operators-2vfzc" Dec 06 03:36:42 crc kubenswrapper[4980]: I1206 03:36:42.242592 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qw2fb\" (UID: \"b25ba819-6e4e-4ca0-a570-c2726b8fbec5\") " pod="openshift-image-registry/image-registry-697d97f7c8-qw2fb" Dec 06 03:36:42 crc kubenswrapper[4980]: E1206 03:36:42.242977 4980 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 03:36:42.742965482 +0000 UTC m=+161.987716753 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qw2fb" (UID: "b25ba819-6e4e-4ca0-a570-c2726b8fbec5") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 03:36:42 crc kubenswrapper[4980]: I1206 03:36:42.243613 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/652a7feb-30e3-4e9f-b1a9-c8ef201c658c-catalog-content\") pod \"community-operators-2vfzc\" (UID: \"652a7feb-30e3-4e9f-b1a9-c8ef201c658c\") " pod="openshift-marketplace/community-operators-2vfzc" Dec 06 03:36:42 crc kubenswrapper[4980]: I1206 03:36:42.244066 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/652a7feb-30e3-4e9f-b1a9-c8ef201c658c-utilities\") pod \"community-operators-2vfzc\" (UID: \"652a7feb-30e3-4e9f-b1a9-c8ef201c658c\") " pod="openshift-marketplace/community-operators-2vfzc" Dec 06 03:36:42 crc kubenswrapper[4980]: I1206 03:36:42.263960 4980 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-kp68b"] Dec 06 03:36:42 crc kubenswrapper[4980]: I1206 03:36:42.264852 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-kp68b" Dec 06 03:36:42 crc kubenswrapper[4980]: I1206 03:36:42.282578 4980 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-kp68b"] Dec 06 03:36:42 crc kubenswrapper[4980]: I1206 03:36:42.284574 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wxr7g\" (UniqueName: \"kubernetes.io/projected/652a7feb-30e3-4e9f-b1a9-c8ef201c658c-kube-api-access-wxr7g\") pod \"community-operators-2vfzc\" (UID: \"652a7feb-30e3-4e9f-b1a9-c8ef201c658c\") " pod="openshift-marketplace/community-operators-2vfzc" Dec 06 03:36:42 crc kubenswrapper[4980]: I1206 03:36:42.311019 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-mkd2m" Dec 06 03:36:42 crc kubenswrapper[4980]: I1206 03:36:42.363455 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 03:36:42 crc kubenswrapper[4980]: I1206 03:36:42.363854 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-58d94" Dec 06 03:36:42 crc kubenswrapper[4980]: E1206 03:36:42.364368 4980 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 03:36:42.864353836 +0000 UTC m=+162.109105107 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 03:36:42 crc kubenswrapper[4980]: I1206 03:36:42.364624 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-2vfzc" Dec 06 03:36:42 crc kubenswrapper[4980]: I1206 03:36:42.431841 4980 patch_prober.go:28] interesting pod/router-default-5444994796-pzzkd container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 06 03:36:42 crc kubenswrapper[4980]: [-]has-synced failed: reason withheld Dec 06 03:36:42 crc kubenswrapper[4980]: [+]process-running ok Dec 06 03:36:42 crc kubenswrapper[4980]: healthz check failed Dec 06 03:36:42 crc kubenswrapper[4980]: I1206 03:36:42.431940 4980 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-pzzkd" podUID="acb53ac6-3f09-424c-a830-a1e15eec0d62" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 06 03:36:42 crc kubenswrapper[4980]: I1206 03:36:42.469585 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/769abfb5-c18a-4cf0-a033-95b3a808867b-catalog-content\") pod \"certified-operators-kp68b\" (UID: \"769abfb5-c18a-4cf0-a033-95b3a808867b\") " pod="openshift-marketplace/certified-operators-kp68b" Dec 06 03:36:42 crc kubenswrapper[4980]: I1206 03:36:42.470007 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/769abfb5-c18a-4cf0-a033-95b3a808867b-utilities\") pod \"certified-operators-kp68b\" (UID: \"769abfb5-c18a-4cf0-a033-95b3a808867b\") " pod="openshift-marketplace/certified-operators-kp68b" Dec 06 03:36:42 crc kubenswrapper[4980]: I1206 03:36:42.470052 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qw2fb\" (UID: \"b25ba819-6e4e-4ca0-a570-c2726b8fbec5\") " pod="openshift-image-registry/image-registry-697d97f7c8-qw2fb" Dec 06 03:36:42 crc kubenswrapper[4980]: I1206 03:36:42.470094 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cnnh8\" (UniqueName: \"kubernetes.io/projected/769abfb5-c18a-4cf0-a033-95b3a808867b-kube-api-access-cnnh8\") pod \"certified-operators-kp68b\" (UID: \"769abfb5-c18a-4cf0-a033-95b3a808867b\") " pod="openshift-marketplace/certified-operators-kp68b" Dec 06 03:36:42 crc kubenswrapper[4980]: E1206 03:36:42.470548 4980 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 03:36:42.970532711 +0000 UTC m=+162.215283982 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qw2fb" (UID: "b25ba819-6e4e-4ca0-a570-c2726b8fbec5") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 03:36:42 crc kubenswrapper[4980]: I1206 03:36:42.566203 4980 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-admission-controller-857f4d67dd-tlpd8" podStartSLOduration=139.566182363 podStartE2EDuration="2m19.566182363s" podCreationTimestamp="2025-12-06 03:34:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 03:36:42.565646617 +0000 UTC m=+161.810397888" watchObservedRunningTime="2025-12-06 03:36:42.566182363 +0000 UTC m=+161.810933634" Dec 06 03:36:42 crc kubenswrapper[4980]: I1206 03:36:42.571056 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 03:36:42 crc kubenswrapper[4980]: I1206 03:36:42.571262 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/769abfb5-c18a-4cf0-a033-95b3a808867b-utilities\") pod \"certified-operators-kp68b\" (UID: \"769abfb5-c18a-4cf0-a033-95b3a808867b\") " pod="openshift-marketplace/certified-operators-kp68b" Dec 06 03:36:42 crc kubenswrapper[4980]: E1206 03:36:42.571370 4980 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 03:36:43.071344262 +0000 UTC m=+162.316095533 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 03:36:42 crc kubenswrapper[4980]: I1206 03:36:42.571431 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cnnh8\" (UniqueName: \"kubernetes.io/projected/769abfb5-c18a-4cf0-a033-95b3a808867b-kube-api-access-cnnh8\") pod \"certified-operators-kp68b\" (UID: \"769abfb5-c18a-4cf0-a033-95b3a808867b\") " pod="openshift-marketplace/certified-operators-kp68b" Dec 06 03:36:42 crc kubenswrapper[4980]: I1206 03:36:42.571577 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/769abfb5-c18a-4cf0-a033-95b3a808867b-catalog-content\") pod \"certified-operators-kp68b\" (UID: \"769abfb5-c18a-4cf0-a033-95b3a808867b\") " pod="openshift-marketplace/certified-operators-kp68b" Dec 06 03:36:42 crc kubenswrapper[4980]: I1206 03:36:42.571827 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/769abfb5-c18a-4cf0-a033-95b3a808867b-utilities\") pod \"certified-operators-kp68b\" (UID: \"769abfb5-c18a-4cf0-a033-95b3a808867b\") " pod="openshift-marketplace/certified-operators-kp68b" Dec 06 03:36:42 crc kubenswrapper[4980]: I1206 03:36:42.572034 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/769abfb5-c18a-4cf0-a033-95b3a808867b-catalog-content\") pod \"certified-operators-kp68b\" (UID: \"769abfb5-c18a-4cf0-a033-95b3a808867b\") " pod="openshift-marketplace/certified-operators-kp68b" Dec 06 03:36:42 crc kubenswrapper[4980]: I1206 03:36:42.609274 4980 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-d6dd4" podStartSLOduration=139.609252236 podStartE2EDuration="2m19.609252236s" podCreationTimestamp="2025-12-06 03:34:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 03:36:42.604023675 +0000 UTC m=+161.848774946" watchObservedRunningTime="2025-12-06 03:36:42.609252236 +0000 UTC m=+161.854003497" Dec 06 03:36:42 crc kubenswrapper[4980]: I1206 03:36:42.619424 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-4dcjh" event={"ID":"8be432d7-9e4b-4452-81df-91998326f92b","Type":"ContainerStarted","Data":"e550ccec5f7b2558e1265b06fb4b9b32455c68cd7950cededa5362e237b0576f"} Dec 06 03:36:42 crc kubenswrapper[4980]: I1206 03:36:42.619631 4980 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-dns/dns-default-4dcjh" Dec 06 03:36:42 crc kubenswrapper[4980]: I1206 03:36:42.622466 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cnnh8\" (UniqueName: \"kubernetes.io/projected/769abfb5-c18a-4cf0-a033-95b3a808867b-kube-api-access-cnnh8\") pod \"certified-operators-kp68b\" (UID: \"769abfb5-c18a-4cf0-a033-95b3a808867b\") " pod="openshift-marketplace/certified-operators-kp68b" Dec 06 03:36:42 crc kubenswrapper[4980]: I1206 03:36:42.651567 4980 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-spd2v" podStartSLOduration=139.651544867 podStartE2EDuration="2m19.651544867s" podCreationTimestamp="2025-12-06 03:34:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 03:36:42.650090715 +0000 UTC m=+161.894841986" watchObservedRunningTime="2025-12-06 03:36:42.651544867 +0000 UTC m=+161.896296138" Dec 06 03:36:42 crc kubenswrapper[4980]: I1206 03:36:42.656858 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-s6h9g" event={"ID":"2cf31abb-ef18-4ba5-8cc9-bf3f7507f771","Type":"ContainerStarted","Data":"10cdc94bd7d5f826310e06b0ab0b33d8c861730b30a669d7fc617cfcc1ebaad4"} Dec 06 03:36:42 crc kubenswrapper[4980]: I1206 03:36:42.657475 4980 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-spd2v container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.23:8080/healthz\": dial tcp 10.217.0.23:8080: connect: connection refused" start-of-body= Dec 06 03:36:42 crc kubenswrapper[4980]: I1206 03:36:42.657541 4980 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-spd2v" podUID="0a45bc54-be9c-4c4f-9cae-dc369d25be33" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.23:8080/healthz\": dial tcp 10.217.0.23:8080: connect: connection refused" Dec 06 03:36:42 crc kubenswrapper[4980]: I1206 03:36:42.657957 4980 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-gjdsq container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.43:5443/healthz\": dial tcp 10.217.0.43:5443: connect: connection refused" start-of-body= Dec 06 03:36:42 crc kubenswrapper[4980]: I1206 03:36:42.658000 4980 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-gjdsq" podUID="949f499a-5680-4fec-ab59-5ced4f24ba2f" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.43:5443/healthz\": dial tcp 10.217.0.43:5443: connect: connection refused" Dec 06 03:36:42 crc kubenswrapper[4980]: I1206 03:36:42.675569 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qw2fb\" (UID: \"b25ba819-6e4e-4ca0-a570-c2726b8fbec5\") " pod="openshift-image-registry/image-registry-697d97f7c8-qw2fb" Dec 06 03:36:42 crc kubenswrapper[4980]: E1206 03:36:42.675943 4980 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 03:36:43.175928011 +0000 UTC m=+162.420679282 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qw2fb" (UID: "b25ba819-6e4e-4ca0-a570-c2726b8fbec5") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 03:36:42 crc kubenswrapper[4980]: I1206 03:36:42.701497 4980 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-qcn9f" podStartSLOduration=139.701481789 podStartE2EDuration="2m19.701481789s" podCreationTimestamp="2025-12-06 03:34:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 03:36:42.699856312 +0000 UTC m=+161.944607573" watchObservedRunningTime="2025-12-06 03:36:42.701481789 +0000 UTC m=+161.946233060" Dec 06 03:36:42 crc kubenswrapper[4980]: I1206 03:36:42.711082 4980 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-sq6bp" Dec 06 03:36:42 crc kubenswrapper[4980]: I1206 03:36:42.770877 4980 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-27gc7" podStartSLOduration=139.770860172 podStartE2EDuration="2m19.770860172s" podCreationTimestamp="2025-12-06 03:34:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 03:36:42.769001018 +0000 UTC m=+162.013752289" watchObservedRunningTime="2025-12-06 03:36:42.770860172 +0000 UTC m=+162.015611443" Dec 06 03:36:42 crc kubenswrapper[4980]: I1206 03:36:42.776923 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 03:36:42 crc kubenswrapper[4980]: E1206 03:36:42.778200 4980 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 03:36:43.278183443 +0000 UTC m=+162.522934714 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 03:36:42 crc kubenswrapper[4980]: I1206 03:36:42.879341 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qw2fb\" (UID: \"b25ba819-6e4e-4ca0-a570-c2726b8fbec5\") " pod="openshift-image-registry/image-registry-697d97f7c8-qw2fb" Dec 06 03:36:42 crc kubenswrapper[4980]: E1206 03:36:42.879673 4980 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 03:36:43.379662433 +0000 UTC m=+162.624413704 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qw2fb" (UID: "b25ba819-6e4e-4ca0-a570-c2726b8fbec5") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 03:36:42 crc kubenswrapper[4980]: I1206 03:36:42.913539 4980 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-f9d7485db-gjmj8" Dec 06 03:36:42 crc kubenswrapper[4980]: I1206 03:36:42.913587 4980 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-f9d7485db-gjmj8" Dec 06 03:36:42 crc kubenswrapper[4980]: I1206 03:36:42.915777 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-kp68b" Dec 06 03:36:42 crc kubenswrapper[4980]: I1206 03:36:42.920618 4980 patch_prober.go:28] interesting pod/console-f9d7485db-gjmj8 container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.6:8443/health\": dial tcp 10.217.0.6:8443: connect: connection refused" start-of-body= Dec 06 03:36:42 crc kubenswrapper[4980]: I1206 03:36:42.920671 4980 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-gjmj8" podUID="443a2dc9-0546-4ab0-a448-eab81c09b482" containerName="console" probeResult="failure" output="Get \"https://10.217.0.6:8443/health\": dial tcp 10.217.0.6:8443: connect: connection refused" Dec 06 03:36:42 crc kubenswrapper[4980]: I1206 03:36:42.921087 4980 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-sq6bp" podStartSLOduration=139.921071058 podStartE2EDuration="2m19.921071058s" podCreationTimestamp="2025-12-06 03:34:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 03:36:42.804201504 +0000 UTC m=+162.048952775" watchObservedRunningTime="2025-12-06 03:36:42.921071058 +0000 UTC m=+162.165822329" Dec 06 03:36:42 crc kubenswrapper[4980]: I1206 03:36:42.956122 4980 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-jp25g" podStartSLOduration=139.95609891 podStartE2EDuration="2m19.95609891s" podCreationTimestamp="2025-12-06 03:34:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 03:36:42.922628583 +0000 UTC m=+162.167379854" watchObservedRunningTime="2025-12-06 03:36:42.95609891 +0000 UTC m=+162.200850181" Dec 06 03:36:42 crc kubenswrapper[4980]: I1206 03:36:42.980177 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 03:36:42 crc kubenswrapper[4980]: E1206 03:36:42.980537 4980 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 03:36:43.480522585 +0000 UTC m=+162.725273856 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 03:36:42 crc kubenswrapper[4980]: I1206 03:36:42.997565 4980 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca/service-ca-9c57cc56f-qpf6b" podStartSLOduration=139.997542626 podStartE2EDuration="2m19.997542626s" podCreationTimestamp="2025-12-06 03:34:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 03:36:42.957881211 +0000 UTC m=+162.202632482" watchObservedRunningTime="2025-12-06 03:36:42.997542626 +0000 UTC m=+162.242293897" Dec 06 03:36:43 crc kubenswrapper[4980]: I1206 03:36:43.027324 4980 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-cjvdn" podStartSLOduration=140.027300925 podStartE2EDuration="2m20.027300925s" podCreationTimestamp="2025-12-06 03:34:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 03:36:42.996418003 +0000 UTC m=+162.241169274" watchObservedRunningTime="2025-12-06 03:36:43.027300925 +0000 UTC m=+162.272052196" Dec 06 03:36:43 crc kubenswrapper[4980]: I1206 03:36:43.098569 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qw2fb\" (UID: \"b25ba819-6e4e-4ca0-a570-c2726b8fbec5\") " pod="openshift-image-registry/image-registry-697d97f7c8-qw2fb" Dec 06 03:36:43 crc kubenswrapper[4980]: E1206 03:36:43.098989 4980 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 03:36:43.598973584 +0000 UTC m=+162.843724855 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qw2fb" (UID: "b25ba819-6e4e-4ca0-a570-c2726b8fbec5") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 03:36:43 crc kubenswrapper[4980]: I1206 03:36:43.104191 4980 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-56s8p" podStartSLOduration=140.104174704 podStartE2EDuration="2m20.104174704s" podCreationTimestamp="2025-12-06 03:34:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 03:36:43.099823559 +0000 UTC m=+162.344574830" watchObservedRunningTime="2025-12-06 03:36:43.104174704 +0000 UTC m=+162.348925975" Dec 06 03:36:43 crc kubenswrapper[4980]: I1206 03:36:43.126762 4980 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-49749" podStartSLOduration=140.126741326 podStartE2EDuration="2m20.126741326s" podCreationTimestamp="2025-12-06 03:34:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 03:36:43.126569481 +0000 UTC m=+162.371320752" watchObservedRunningTime="2025-12-06 03:36:43.126741326 +0000 UTC m=+162.371492607" Dec 06 03:36:43 crc kubenswrapper[4980]: I1206 03:36:43.139073 4980 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-66bs5" Dec 06 03:36:43 crc kubenswrapper[4980]: I1206 03:36:43.142729 4980 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-66bs5" Dec 06 03:36:43 crc kubenswrapper[4980]: I1206 03:36:43.180313 4980 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-gjdsq" podStartSLOduration=140.180295142 podStartE2EDuration="2m20.180295142s" podCreationTimestamp="2025-12-06 03:34:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 03:36:43.178878191 +0000 UTC m=+162.423629462" watchObservedRunningTime="2025-12-06 03:36:43.180295142 +0000 UTC m=+162.425046413" Dec 06 03:36:43 crc kubenswrapper[4980]: I1206 03:36:43.181870 4980 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-66bs5" Dec 06 03:36:43 crc kubenswrapper[4980]: I1206 03:36:43.206319 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 03:36:43 crc kubenswrapper[4980]: E1206 03:36:43.206959 4980 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 03:36:43.706941881 +0000 UTC m=+162.951693152 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 03:36:43 crc kubenswrapper[4980]: I1206 03:36:43.303225 4980 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns-operator/dns-operator-744455d44c-w6bmg" podStartSLOduration=140.30320412 podStartE2EDuration="2m20.30320412s" podCreationTimestamp="2025-12-06 03:34:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 03:36:43.233802966 +0000 UTC m=+162.478554247" watchObservedRunningTime="2025-12-06 03:36:43.30320412 +0000 UTC m=+162.547955391" Dec 06 03:36:43 crc kubenswrapper[4980]: I1206 03:36:43.307727 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qw2fb\" (UID: \"b25ba819-6e4e-4ca0-a570-c2726b8fbec5\") " pod="openshift-image-registry/image-registry-697d97f7c8-qw2fb" Dec 06 03:36:43 crc kubenswrapper[4980]: E1206 03:36:43.308036 4980 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 03:36:43.808023049 +0000 UTC m=+163.052774320 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qw2fb" (UID: "b25ba819-6e4e-4ca0-a570-c2726b8fbec5") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 03:36:43 crc kubenswrapper[4980]: I1206 03:36:43.416217 4980 patch_prober.go:28] interesting pod/router-default-5444994796-pzzkd container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 06 03:36:43 crc kubenswrapper[4980]: [-]has-synced failed: reason withheld Dec 06 03:36:43 crc kubenswrapper[4980]: [+]process-running ok Dec 06 03:36:43 crc kubenswrapper[4980]: healthz check failed Dec 06 03:36:43 crc kubenswrapper[4980]: I1206 03:36:43.416303 4980 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-pzzkd" podUID="acb53ac6-3f09-424c-a830-a1e15eec0d62" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 06 03:36:43 crc kubenswrapper[4980]: I1206 03:36:43.425852 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 03:36:43 crc kubenswrapper[4980]: E1206 03:36:43.426164 4980 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 03:36:43.926149439 +0000 UTC m=+163.170900700 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 03:36:43 crc kubenswrapper[4980]: I1206 03:36:43.482823 4980 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/dns-default-4dcjh" podStartSLOduration=12.482800895 podStartE2EDuration="12.482800895s" podCreationTimestamp="2025-12-06 03:36:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 03:36:43.479533781 +0000 UTC m=+162.724285052" watchObservedRunningTime="2025-12-06 03:36:43.482800895 +0000 UTC m=+162.727552176" Dec 06 03:36:43 crc kubenswrapper[4980]: I1206 03:36:43.538066 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qw2fb\" (UID: \"b25ba819-6e4e-4ca0-a570-c2726b8fbec5\") " pod="openshift-image-registry/image-registry-697d97f7c8-qw2fb" Dec 06 03:36:43 crc kubenswrapper[4980]: E1206 03:36:43.538351 4980 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 03:36:44.038338468 +0000 UTC m=+163.283089739 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qw2fb" (UID: "b25ba819-6e4e-4ca0-a570-c2726b8fbec5") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 03:36:43 crc kubenswrapper[4980]: I1206 03:36:43.539571 4980 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver/apiserver-76f77b778f-s6h9g" podStartSLOduration=141.539562824 podStartE2EDuration="2m21.539562824s" podCreationTimestamp="2025-12-06 03:34:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 03:36:43.538856903 +0000 UTC m=+162.783608174" watchObservedRunningTime="2025-12-06 03:36:43.539562824 +0000 UTC m=+162.784314095" Dec 06 03:36:43 crc kubenswrapper[4980]: I1206 03:36:43.613768 4980 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-58d94"] Dec 06 03:36:43 crc kubenswrapper[4980]: I1206 03:36:43.665537 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 03:36:43 crc kubenswrapper[4980]: E1206 03:36:43.665964 4980 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 03:36:44.165944272 +0000 UTC m=+163.410695543 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 03:36:43 crc kubenswrapper[4980]: I1206 03:36:43.718317 4980 generic.go:334] "Generic (PLEG): container finished" podID="03704eb9-227d-4985-87cf-acd66ae76037" containerID="c4afa62f429f5f866be3d2100de061f502a4becc8ee55d85f991a5589586415c" exitCode=0 Dec 06 03:36:43 crc kubenswrapper[4980]: I1206 03:36:43.718397 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29416530-9hm9m" event={"ID":"03704eb9-227d-4985-87cf-acd66ae76037","Type":"ContainerDied","Data":"c4afa62f429f5f866be3d2100de061f502a4becc8ee55d85f991a5589586415c"} Dec 06 03:36:43 crc kubenswrapper[4980]: I1206 03:36:43.733761 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-74hx2" event={"ID":"9b9bddd5-f192-4b4c-81ee-b523bf9c8705","Type":"ContainerStarted","Data":"063771e811c098e0b8d914b71a4a55f59478c1206e93c85bc40f78436fb8b219"} Dec 06 03:36:43 crc kubenswrapper[4980]: I1206 03:36:43.735096 4980 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-spd2v container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.23:8080/healthz\": dial tcp 10.217.0.23:8080: connect: connection refused" start-of-body= Dec 06 03:36:43 crc kubenswrapper[4980]: I1206 03:36:43.735151 4980 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-spd2v" podUID="0a45bc54-be9c-4c4f-9cae-dc369d25be33" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.23:8080/healthz\": dial tcp 10.217.0.23:8080: connect: connection refused" Dec 06 03:36:43 crc kubenswrapper[4980]: I1206 03:36:43.747547 4980 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-66bs5" Dec 06 03:36:43 crc kubenswrapper[4980]: I1206 03:36:43.768002 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qw2fb\" (UID: \"b25ba819-6e4e-4ca0-a570-c2726b8fbec5\") " pod="openshift-image-registry/image-registry-697d97f7c8-qw2fb" Dec 06 03:36:43 crc kubenswrapper[4980]: E1206 03:36:43.768365 4980 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 03:36:44.268353969 +0000 UTC m=+163.513105240 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qw2fb" (UID: "b25ba819-6e4e-4ca0-a570-c2726b8fbec5") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 03:36:43 crc kubenswrapper[4980]: I1206 03:36:43.812976 4980 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-2vfzc"] Dec 06 03:36:43 crc kubenswrapper[4980]: I1206 03:36:43.813145 4980 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-kp68b"] Dec 06 03:36:43 crc kubenswrapper[4980]: I1206 03:36:43.822823 4980 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-mkd2m"] Dec 06 03:36:43 crc kubenswrapper[4980]: I1206 03:36:43.872490 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 03:36:43 crc kubenswrapper[4980]: E1206 03:36:43.873942 4980 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 03:36:44.373908736 +0000 UTC m=+163.618660177 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 03:36:43 crc kubenswrapper[4980]: I1206 03:36:43.875767 4980 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-gxcqb"] Dec 06 03:36:43 crc kubenswrapper[4980]: I1206 03:36:43.877204 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-gxcqb" Dec 06 03:36:43 crc kubenswrapper[4980]: I1206 03:36:43.896368 4980 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Dec 06 03:36:43 crc kubenswrapper[4980]: I1206 03:36:43.898058 4980 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-gxcqb"] Dec 06 03:36:43 crc kubenswrapper[4980]: I1206 03:36:43.979095 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b9kxc\" (UniqueName: \"kubernetes.io/projected/2ce9bc16-3d2a-4feb-82c1-0f42518196cf-kube-api-access-b9kxc\") pod \"redhat-marketplace-gxcqb\" (UID: \"2ce9bc16-3d2a-4feb-82c1-0f42518196cf\") " pod="openshift-marketplace/redhat-marketplace-gxcqb" Dec 06 03:36:43 crc kubenswrapper[4980]: I1206 03:36:43.979161 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qw2fb\" (UID: \"b25ba819-6e4e-4ca0-a570-c2726b8fbec5\") " pod="openshift-image-registry/image-registry-697d97f7c8-qw2fb" Dec 06 03:36:43 crc kubenswrapper[4980]: I1206 03:36:43.979212 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2ce9bc16-3d2a-4feb-82c1-0f42518196cf-catalog-content\") pod \"redhat-marketplace-gxcqb\" (UID: \"2ce9bc16-3d2a-4feb-82c1-0f42518196cf\") " pod="openshift-marketplace/redhat-marketplace-gxcqb" Dec 06 03:36:43 crc kubenswrapper[4980]: I1206 03:36:43.979250 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2ce9bc16-3d2a-4feb-82c1-0f42518196cf-utilities\") pod \"redhat-marketplace-gxcqb\" (UID: \"2ce9bc16-3d2a-4feb-82c1-0f42518196cf\") " pod="openshift-marketplace/redhat-marketplace-gxcqb" Dec 06 03:36:43 crc kubenswrapper[4980]: E1206 03:36:43.995054 4980 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 03:36:44.495034673 +0000 UTC m=+163.739785944 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qw2fb" (UID: "b25ba819-6e4e-4ca0-a570-c2726b8fbec5") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 03:36:44 crc kubenswrapper[4980]: I1206 03:36:44.026117 4980 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-apiserver/apiserver-76f77b778f-s6h9g" Dec 06 03:36:44 crc kubenswrapper[4980]: I1206 03:36:44.026227 4980 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-apiserver/apiserver-76f77b778f-s6h9g" Dec 06 03:36:44 crc kubenswrapper[4980]: I1206 03:36:44.033630 4980 patch_prober.go:28] interesting pod/apiserver-76f77b778f-s6h9g container/openshift-apiserver namespace/openshift-apiserver: Startup probe status=failure output="Get \"https://10.217.0.5:8443/livez\": dial tcp 10.217.0.5:8443: connect: connection refused" start-of-body= Dec 06 03:36:44 crc kubenswrapper[4980]: I1206 03:36:44.033694 4980 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-apiserver/apiserver-76f77b778f-s6h9g" podUID="2cf31abb-ef18-4ba5-8cc9-bf3f7507f771" containerName="openshift-apiserver" probeResult="failure" output="Get \"https://10.217.0.5:8443/livez\": dial tcp 10.217.0.5:8443: connect: connection refused" Dec 06 03:36:44 crc kubenswrapper[4980]: I1206 03:36:44.088185 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 03:36:44 crc kubenswrapper[4980]: I1206 03:36:44.088756 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2ce9bc16-3d2a-4feb-82c1-0f42518196cf-catalog-content\") pod \"redhat-marketplace-gxcqb\" (UID: \"2ce9bc16-3d2a-4feb-82c1-0f42518196cf\") " pod="openshift-marketplace/redhat-marketplace-gxcqb" Dec 06 03:36:44 crc kubenswrapper[4980]: I1206 03:36:44.088803 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2ce9bc16-3d2a-4feb-82c1-0f42518196cf-utilities\") pod \"redhat-marketplace-gxcqb\" (UID: \"2ce9bc16-3d2a-4feb-82c1-0f42518196cf\") " pod="openshift-marketplace/redhat-marketplace-gxcqb" Dec 06 03:36:44 crc kubenswrapper[4980]: I1206 03:36:44.088889 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b9kxc\" (UniqueName: \"kubernetes.io/projected/2ce9bc16-3d2a-4feb-82c1-0f42518196cf-kube-api-access-b9kxc\") pod \"redhat-marketplace-gxcqb\" (UID: \"2ce9bc16-3d2a-4feb-82c1-0f42518196cf\") " pod="openshift-marketplace/redhat-marketplace-gxcqb" Dec 06 03:36:44 crc kubenswrapper[4980]: E1206 03:36:44.089422 4980 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 03:36:44.589406398 +0000 UTC m=+163.834157669 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 03:36:44 crc kubenswrapper[4980]: I1206 03:36:44.089996 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2ce9bc16-3d2a-4feb-82c1-0f42518196cf-catalog-content\") pod \"redhat-marketplace-gxcqb\" (UID: \"2ce9bc16-3d2a-4feb-82c1-0f42518196cf\") " pod="openshift-marketplace/redhat-marketplace-gxcqb" Dec 06 03:36:44 crc kubenswrapper[4980]: I1206 03:36:44.090272 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2ce9bc16-3d2a-4feb-82c1-0f42518196cf-utilities\") pod \"redhat-marketplace-gxcqb\" (UID: \"2ce9bc16-3d2a-4feb-82c1-0f42518196cf\") " pod="openshift-marketplace/redhat-marketplace-gxcqb" Dec 06 03:36:44 crc kubenswrapper[4980]: I1206 03:36:44.118560 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b9kxc\" (UniqueName: \"kubernetes.io/projected/2ce9bc16-3d2a-4feb-82c1-0f42518196cf-kube-api-access-b9kxc\") pod \"redhat-marketplace-gxcqb\" (UID: \"2ce9bc16-3d2a-4feb-82c1-0f42518196cf\") " pod="openshift-marketplace/redhat-marketplace-gxcqb" Dec 06 03:36:44 crc kubenswrapper[4980]: I1206 03:36:44.189970 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qw2fb\" (UID: \"b25ba819-6e4e-4ca0-a570-c2726b8fbec5\") " pod="openshift-image-registry/image-registry-697d97f7c8-qw2fb" Dec 06 03:36:44 crc kubenswrapper[4980]: E1206 03:36:44.190380 4980 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 03:36:44.690364162 +0000 UTC m=+163.935115433 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qw2fb" (UID: "b25ba819-6e4e-4ca0-a570-c2726b8fbec5") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 03:36:44 crc kubenswrapper[4980]: I1206 03:36:44.214487 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-gxcqb" Dec 06 03:36:44 crc kubenswrapper[4980]: I1206 03:36:44.238181 4980 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-xhnw8"] Dec 06 03:36:44 crc kubenswrapper[4980]: I1206 03:36:44.239113 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-xhnw8" Dec 06 03:36:44 crc kubenswrapper[4980]: I1206 03:36:44.251166 4980 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-xhnw8"] Dec 06 03:36:44 crc kubenswrapper[4980]: I1206 03:36:44.290746 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 03:36:44 crc kubenswrapper[4980]: E1206 03:36:44.291092 4980 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 03:36:44.7910784 +0000 UTC m=+164.035829671 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 03:36:44 crc kubenswrapper[4980]: I1206 03:36:44.297817 4980 patch_prober.go:28] interesting pod/downloads-7954f5f757-6h96g container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.9:8080/\": dial tcp 10.217.0.9:8080: connect: connection refused" start-of-body= Dec 06 03:36:44 crc kubenswrapper[4980]: I1206 03:36:44.297860 4980 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-6h96g" podUID="346bb011-c34e-4286-90d3-0242205e8e94" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.9:8080/\": dial tcp 10.217.0.9:8080: connect: connection refused" Dec 06 03:36:44 crc kubenswrapper[4980]: I1206 03:36:44.297911 4980 patch_prober.go:28] interesting pod/downloads-7954f5f757-6h96g container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.9:8080/\": dial tcp 10.217.0.9:8080: connect: connection refused" start-of-body= Dec 06 03:36:44 crc kubenswrapper[4980]: I1206 03:36:44.297925 4980 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-6h96g" podUID="346bb011-c34e-4286-90d3-0242205e8e94" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.9:8080/\": dial tcp 10.217.0.9:8080: connect: connection refused" Dec 06 03:36:44 crc kubenswrapper[4980]: I1206 03:36:44.393149 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zgc7q\" (UniqueName: \"kubernetes.io/projected/3caa51b4-a311-4498-ac9b-10c20989684a-kube-api-access-zgc7q\") pod \"redhat-marketplace-xhnw8\" (UID: \"3caa51b4-a311-4498-ac9b-10c20989684a\") " pod="openshift-marketplace/redhat-marketplace-xhnw8" Dec 06 03:36:44 crc kubenswrapper[4980]: I1206 03:36:44.393652 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3caa51b4-a311-4498-ac9b-10c20989684a-utilities\") pod \"redhat-marketplace-xhnw8\" (UID: \"3caa51b4-a311-4498-ac9b-10c20989684a\") " pod="openshift-marketplace/redhat-marketplace-xhnw8" Dec 06 03:36:44 crc kubenswrapper[4980]: I1206 03:36:44.393780 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qw2fb\" (UID: \"b25ba819-6e4e-4ca0-a570-c2726b8fbec5\") " pod="openshift-image-registry/image-registry-697d97f7c8-qw2fb" Dec 06 03:36:44 crc kubenswrapper[4980]: I1206 03:36:44.393809 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3caa51b4-a311-4498-ac9b-10c20989684a-catalog-content\") pod \"redhat-marketplace-xhnw8\" (UID: \"3caa51b4-a311-4498-ac9b-10c20989684a\") " pod="openshift-marketplace/redhat-marketplace-xhnw8" Dec 06 03:36:44 crc kubenswrapper[4980]: E1206 03:36:44.394190 4980 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 03:36:44.894176306 +0000 UTC m=+164.138927577 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qw2fb" (UID: "b25ba819-6e4e-4ca0-a570-c2726b8fbec5") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 03:36:44 crc kubenswrapper[4980]: I1206 03:36:44.413780 4980 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ingress/router-default-5444994796-pzzkd" Dec 06 03:36:44 crc kubenswrapper[4980]: I1206 03:36:44.427103 4980 patch_prober.go:28] interesting pod/router-default-5444994796-pzzkd container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 06 03:36:44 crc kubenswrapper[4980]: [-]has-synced failed: reason withheld Dec 06 03:36:44 crc kubenswrapper[4980]: [+]process-running ok Dec 06 03:36:44 crc kubenswrapper[4980]: healthz check failed Dec 06 03:36:44 crc kubenswrapper[4980]: I1206 03:36:44.427155 4980 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-pzzkd" podUID="acb53ac6-3f09-424c-a830-a1e15eec0d62" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 06 03:36:44 crc kubenswrapper[4980]: I1206 03:36:44.481085 4980 plugin_watcher.go:194] "Adding socket path or updating timestamp to desired state cache" path="/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock" Dec 06 03:36:44 crc kubenswrapper[4980]: I1206 03:36:44.495356 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 03:36:44 crc kubenswrapper[4980]: I1206 03:36:44.495834 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zgc7q\" (UniqueName: \"kubernetes.io/projected/3caa51b4-a311-4498-ac9b-10c20989684a-kube-api-access-zgc7q\") pod \"redhat-marketplace-xhnw8\" (UID: \"3caa51b4-a311-4498-ac9b-10c20989684a\") " pod="openshift-marketplace/redhat-marketplace-xhnw8" Dec 06 03:36:44 crc kubenswrapper[4980]: I1206 03:36:44.495995 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3caa51b4-a311-4498-ac9b-10c20989684a-utilities\") pod \"redhat-marketplace-xhnw8\" (UID: \"3caa51b4-a311-4498-ac9b-10c20989684a\") " pod="openshift-marketplace/redhat-marketplace-xhnw8" Dec 06 03:36:44 crc kubenswrapper[4980]: I1206 03:36:44.496368 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3caa51b4-a311-4498-ac9b-10c20989684a-catalog-content\") pod \"redhat-marketplace-xhnw8\" (UID: \"3caa51b4-a311-4498-ac9b-10c20989684a\") " pod="openshift-marketplace/redhat-marketplace-xhnw8" Dec 06 03:36:44 crc kubenswrapper[4980]: E1206 03:36:44.497206 4980 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 03:36:44.99718613 +0000 UTC m=+164.241937431 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 03:36:44 crc kubenswrapper[4980]: I1206 03:36:44.498162 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3caa51b4-a311-4498-ac9b-10c20989684a-utilities\") pod \"redhat-marketplace-xhnw8\" (UID: \"3caa51b4-a311-4498-ac9b-10c20989684a\") " pod="openshift-marketplace/redhat-marketplace-xhnw8" Dec 06 03:36:44 crc kubenswrapper[4980]: I1206 03:36:44.498299 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3caa51b4-a311-4498-ac9b-10c20989684a-catalog-content\") pod \"redhat-marketplace-xhnw8\" (UID: \"3caa51b4-a311-4498-ac9b-10c20989684a\") " pod="openshift-marketplace/redhat-marketplace-xhnw8" Dec 06 03:36:44 crc kubenswrapper[4980]: I1206 03:36:44.550419 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zgc7q\" (UniqueName: \"kubernetes.io/projected/3caa51b4-a311-4498-ac9b-10c20989684a-kube-api-access-zgc7q\") pod \"redhat-marketplace-xhnw8\" (UID: \"3caa51b4-a311-4498-ac9b-10c20989684a\") " pod="openshift-marketplace/redhat-marketplace-xhnw8" Dec 06 03:36:44 crc kubenswrapper[4980]: I1206 03:36:44.552845 4980 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-spd2v" Dec 06 03:36:44 crc kubenswrapper[4980]: I1206 03:36:44.590771 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-xhnw8" Dec 06 03:36:44 crc kubenswrapper[4980]: I1206 03:36:44.602440 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qw2fb\" (UID: \"b25ba819-6e4e-4ca0-a570-c2726b8fbec5\") " pod="openshift-image-registry/image-registry-697d97f7c8-qw2fb" Dec 06 03:36:44 crc kubenswrapper[4980]: E1206 03:36:44.602886 4980 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 03:36:45.102870031 +0000 UTC m=+164.347621302 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qw2fb" (UID: "b25ba819-6e4e-4ca0-a570-c2726b8fbec5") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 03:36:44 crc kubenswrapper[4980]: I1206 03:36:44.693028 4980 reconciler.go:161] "OperationExecutor.RegisterPlugin started" plugin={"SocketPath":"/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock","Timestamp":"2025-12-06T03:36:44.481321562Z","Handler":null,"Name":""} Dec 06 03:36:44 crc kubenswrapper[4980]: I1206 03:36:44.752978 4980 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-gjdsq container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.43:5443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 06 03:36:44 crc kubenswrapper[4980]: I1206 03:36:44.753073 4980 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-gjdsq" podUID="949f499a-5680-4fec-ab59-5ced4f24ba2f" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.43:5443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 06 03:36:44 crc kubenswrapper[4980]: I1206 03:36:44.759399 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 03:36:44 crc kubenswrapper[4980]: E1206 03:36:44.760044 4980 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 03:36:45.259998327 +0000 UTC m=+164.504749598 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 03:36:44 crc kubenswrapper[4980]: I1206 03:36:44.818694 4980 generic.go:334] "Generic (PLEG): container finished" podID="769abfb5-c18a-4cf0-a033-95b3a808867b" containerID="ef2befdf9890e17b7d072571f6b80abbd83962e482172e3dc516ca5c403631c6" exitCode=0 Dec 06 03:36:44 crc kubenswrapper[4980]: I1206 03:36:44.818791 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kp68b" event={"ID":"769abfb5-c18a-4cf0-a033-95b3a808867b","Type":"ContainerDied","Data":"ef2befdf9890e17b7d072571f6b80abbd83962e482172e3dc516ca5c403631c6"} Dec 06 03:36:44 crc kubenswrapper[4980]: I1206 03:36:44.818822 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kp68b" event={"ID":"769abfb5-c18a-4cf0-a033-95b3a808867b","Type":"ContainerStarted","Data":"87b491bed7e9e4c176b0bb2f09b26aa5929116e0c3dd3a76189f34e09567bc1a"} Dec 06 03:36:44 crc kubenswrapper[4980]: I1206 03:36:44.821807 4980 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-gjdsq" Dec 06 03:36:44 crc kubenswrapper[4980]: I1206 03:36:44.824565 4980 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 06 03:36:44 crc kubenswrapper[4980]: I1206 03:36:44.863053 4980 generic.go:334] "Generic (PLEG): container finished" podID="e965bb7a-1bf0-4574-aae4-2a45c75fb11c" containerID="1f375c47d1ab2d4e72c1d6e5a9af8ee6f9280fccbd3ce180afdc0ff7b8a980bf" exitCode=0 Dec 06 03:36:44 crc kubenswrapper[4980]: I1206 03:36:44.863760 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-58d94" event={"ID":"e965bb7a-1bf0-4574-aae4-2a45c75fb11c","Type":"ContainerDied","Data":"1f375c47d1ab2d4e72c1d6e5a9af8ee6f9280fccbd3ce180afdc0ff7b8a980bf"} Dec 06 03:36:44 crc kubenswrapper[4980]: I1206 03:36:44.863807 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-58d94" event={"ID":"e965bb7a-1bf0-4574-aae4-2a45c75fb11c","Type":"ContainerStarted","Data":"5c48468877394b6180088d640646f461561cffea4774804c561aae3e0fda3e81"} Dec 06 03:36:44 crc kubenswrapper[4980]: I1206 03:36:44.863956 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qw2fb\" (UID: \"b25ba819-6e4e-4ca0-a570-c2726b8fbec5\") " pod="openshift-image-registry/image-registry-697d97f7c8-qw2fb" Dec 06 03:36:44 crc kubenswrapper[4980]: E1206 03:36:44.864928 4980 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 03:36:45.364912596 +0000 UTC m=+164.609663867 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qw2fb" (UID: "b25ba819-6e4e-4ca0-a570-c2726b8fbec5") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 03:36:44 crc kubenswrapper[4980]: I1206 03:36:44.875424 4980 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-m9vl6"] Dec 06 03:36:44 crc kubenswrapper[4980]: I1206 03:36:44.876845 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-m9vl6" Dec 06 03:36:44 crc kubenswrapper[4980]: I1206 03:36:44.882740 4980 csi_plugin.go:100] kubernetes.io/csi: Trying to validate a new CSI Driver with name: kubevirt.io.hostpath-provisioner endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock versions: 1.0.0 Dec 06 03:36:44 crc kubenswrapper[4980]: I1206 03:36:44.882875 4980 csi_plugin.go:113] kubernetes.io/csi: Register new plugin with name: kubevirt.io.hostpath-provisioner at endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock Dec 06 03:36:44 crc kubenswrapper[4980]: I1206 03:36:44.887360 4980 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Dec 06 03:36:44 crc kubenswrapper[4980]: I1206 03:36:44.892279 4980 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-m9vl6"] Dec 06 03:36:44 crc kubenswrapper[4980]: I1206 03:36:44.910735 4980 generic.go:334] "Generic (PLEG): container finished" podID="652a7feb-30e3-4e9f-b1a9-c8ef201c658c" containerID="a3b66bc131bf66af908ba3cf9ed3c9d469b481a1db21f645f7785f7051d84f9f" exitCode=0 Dec 06 03:36:44 crc kubenswrapper[4980]: I1206 03:36:44.913956 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2vfzc" event={"ID":"652a7feb-30e3-4e9f-b1a9-c8ef201c658c","Type":"ContainerDied","Data":"a3b66bc131bf66af908ba3cf9ed3c9d469b481a1db21f645f7785f7051d84f9f"} Dec 06 03:36:44 crc kubenswrapper[4980]: I1206 03:36:44.914021 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2vfzc" event={"ID":"652a7feb-30e3-4e9f-b1a9-c8ef201c658c","Type":"ContainerStarted","Data":"4cf3f79b3da4c6177b583e0ef3da23b5cd48f26fa556daa5657f94db53da4a22"} Dec 06 03:36:44 crc kubenswrapper[4980]: I1206 03:36:44.928048 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-74hx2" event={"ID":"9b9bddd5-f192-4b4c-81ee-b523bf9c8705","Type":"ContainerStarted","Data":"ad2e2ad73235f9145acf0734a53f7892ee822b028a13cb8092aed59b591837f5"} Dec 06 03:36:44 crc kubenswrapper[4980]: I1206 03:36:44.944775 4980 generic.go:334] "Generic (PLEG): container finished" podID="9c8b3374-77eb-44ad-b270-2263c6c23e28" containerID="c9d59ae689acb66f2a450b619034ed36904d9c1cc9adfef732f043b58e582bec" exitCode=0 Dec 06 03:36:44 crc kubenswrapper[4980]: I1206 03:36:44.946428 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mkd2m" event={"ID":"9c8b3374-77eb-44ad-b270-2263c6c23e28","Type":"ContainerDied","Data":"c9d59ae689acb66f2a450b619034ed36904d9c1cc9adfef732f043b58e582bec"} Dec 06 03:36:44 crc kubenswrapper[4980]: I1206 03:36:44.946455 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mkd2m" event={"ID":"9c8b3374-77eb-44ad-b270-2263c6c23e28","Type":"ContainerStarted","Data":"48c1a73f4a7ea2d6254b7b0f4f0d5c79560a3376f75717817c4ff9dc9e88b8cd"} Dec 06 03:36:44 crc kubenswrapper[4980]: I1206 03:36:44.965156 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 03:36:44 crc kubenswrapper[4980]: I1206 03:36:44.965582 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hwq2v\" (UniqueName: \"kubernetes.io/projected/43112506-9fdd-4b19-a5a5-caf864c09774-kube-api-access-hwq2v\") pod \"redhat-operators-m9vl6\" (UID: \"43112506-9fdd-4b19-a5a5-caf864c09774\") " pod="openshift-marketplace/redhat-operators-m9vl6" Dec 06 03:36:44 crc kubenswrapper[4980]: I1206 03:36:44.965829 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/43112506-9fdd-4b19-a5a5-caf864c09774-catalog-content\") pod \"redhat-operators-m9vl6\" (UID: \"43112506-9fdd-4b19-a5a5-caf864c09774\") " pod="openshift-marketplace/redhat-operators-m9vl6" Dec 06 03:36:44 crc kubenswrapper[4980]: I1206 03:36:44.965917 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/43112506-9fdd-4b19-a5a5-caf864c09774-utilities\") pod \"redhat-operators-m9vl6\" (UID: \"43112506-9fdd-4b19-a5a5-caf864c09774\") " pod="openshift-marketplace/redhat-operators-m9vl6" Dec 06 03:36:44 crc kubenswrapper[4980]: I1206 03:36:44.990980 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Dec 06 03:36:45 crc kubenswrapper[4980]: I1206 03:36:45.041480 4980 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-gxcqb"] Dec 06 03:36:45 crc kubenswrapper[4980]: I1206 03:36:45.068261 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/43112506-9fdd-4b19-a5a5-caf864c09774-catalog-content\") pod \"redhat-operators-m9vl6\" (UID: \"43112506-9fdd-4b19-a5a5-caf864c09774\") " pod="openshift-marketplace/redhat-operators-m9vl6" Dec 06 03:36:45 crc kubenswrapper[4980]: I1206 03:36:45.068312 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/43112506-9fdd-4b19-a5a5-caf864c09774-utilities\") pod \"redhat-operators-m9vl6\" (UID: \"43112506-9fdd-4b19-a5a5-caf864c09774\") " pod="openshift-marketplace/redhat-operators-m9vl6" Dec 06 03:36:45 crc kubenswrapper[4980]: I1206 03:36:45.068343 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qw2fb\" (UID: \"b25ba819-6e4e-4ca0-a570-c2726b8fbec5\") " pod="openshift-image-registry/image-registry-697d97f7c8-qw2fb" Dec 06 03:36:45 crc kubenswrapper[4980]: I1206 03:36:45.068423 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hwq2v\" (UniqueName: \"kubernetes.io/projected/43112506-9fdd-4b19-a5a5-caf864c09774-kube-api-access-hwq2v\") pod \"redhat-operators-m9vl6\" (UID: \"43112506-9fdd-4b19-a5a5-caf864c09774\") " pod="openshift-marketplace/redhat-operators-m9vl6" Dec 06 03:36:45 crc kubenswrapper[4980]: I1206 03:36:45.069917 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/43112506-9fdd-4b19-a5a5-caf864c09774-utilities\") pod \"redhat-operators-m9vl6\" (UID: \"43112506-9fdd-4b19-a5a5-caf864c09774\") " pod="openshift-marketplace/redhat-operators-m9vl6" Dec 06 03:36:45 crc kubenswrapper[4980]: I1206 03:36:45.070163 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/43112506-9fdd-4b19-a5a5-caf864c09774-catalog-content\") pod \"redhat-operators-m9vl6\" (UID: \"43112506-9fdd-4b19-a5a5-caf864c09774\") " pod="openshift-marketplace/redhat-operators-m9vl6" Dec 06 03:36:45 crc kubenswrapper[4980]: I1206 03:36:45.104872 4980 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 06 03:36:45 crc kubenswrapper[4980]: I1206 03:36:45.105587 4980 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qw2fb\" (UID: \"b25ba819-6e4e-4ca0-a570-c2726b8fbec5\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount\"" pod="openshift-image-registry/image-registry-697d97f7c8-qw2fb" Dec 06 03:36:45 crc kubenswrapper[4980]: I1206 03:36:45.117217 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hwq2v\" (UniqueName: \"kubernetes.io/projected/43112506-9fdd-4b19-a5a5-caf864c09774-kube-api-access-hwq2v\") pod \"redhat-operators-m9vl6\" (UID: \"43112506-9fdd-4b19-a5a5-caf864c09774\") " pod="openshift-marketplace/redhat-operators-m9vl6" Dec 06 03:36:45 crc kubenswrapper[4980]: I1206 03:36:45.224400 4980 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8f668bae-612b-4b75-9490-919e737c6a3b" path="/var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes" Dec 06 03:36:45 crc kubenswrapper[4980]: I1206 03:36:45.242503 4980 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-bl7hx"] Dec 06 03:36:45 crc kubenswrapper[4980]: I1206 03:36:45.243832 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-bl7hx" Dec 06 03:36:45 crc kubenswrapper[4980]: I1206 03:36:45.267440 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-m9vl6" Dec 06 03:36:45 crc kubenswrapper[4980]: I1206 03:36:45.273838 4980 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-bl7hx"] Dec 06 03:36:45 crc kubenswrapper[4980]: I1206 03:36:45.381176 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/56a102c7-f341-4076-89dc-44428e77c164-utilities\") pod \"redhat-operators-bl7hx\" (UID: \"56a102c7-f341-4076-89dc-44428e77c164\") " pod="openshift-marketplace/redhat-operators-bl7hx" Dec 06 03:36:45 crc kubenswrapper[4980]: I1206 03:36:45.381286 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/56a102c7-f341-4076-89dc-44428e77c164-catalog-content\") pod \"redhat-operators-bl7hx\" (UID: \"56a102c7-f341-4076-89dc-44428e77c164\") " pod="openshift-marketplace/redhat-operators-bl7hx" Dec 06 03:36:45 crc kubenswrapper[4980]: I1206 03:36:45.381308 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vpb28\" (UniqueName: \"kubernetes.io/projected/56a102c7-f341-4076-89dc-44428e77c164-kube-api-access-vpb28\") pod \"redhat-operators-bl7hx\" (UID: \"56a102c7-f341-4076-89dc-44428e77c164\") " pod="openshift-marketplace/redhat-operators-bl7hx" Dec 06 03:36:45 crc kubenswrapper[4980]: I1206 03:36:45.432173 4980 patch_prober.go:28] interesting pod/router-default-5444994796-pzzkd container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 06 03:36:45 crc kubenswrapper[4980]: [-]has-synced failed: reason withheld Dec 06 03:36:45 crc kubenswrapper[4980]: [+]process-running ok Dec 06 03:36:45 crc kubenswrapper[4980]: healthz check failed Dec 06 03:36:45 crc kubenswrapper[4980]: I1206 03:36:45.432596 4980 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-pzzkd" podUID="acb53ac6-3f09-424c-a830-a1e15eec0d62" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 06 03:36:45 crc kubenswrapper[4980]: I1206 03:36:45.458670 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qw2fb\" (UID: \"b25ba819-6e4e-4ca0-a570-c2726b8fbec5\") " pod="openshift-image-registry/image-registry-697d97f7c8-qw2fb" Dec 06 03:36:45 crc kubenswrapper[4980]: I1206 03:36:45.478492 4980 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-xhnw8"] Dec 06 03:36:45 crc kubenswrapper[4980]: I1206 03:36:45.482385 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/56a102c7-f341-4076-89dc-44428e77c164-catalog-content\") pod \"redhat-operators-bl7hx\" (UID: \"56a102c7-f341-4076-89dc-44428e77c164\") " pod="openshift-marketplace/redhat-operators-bl7hx" Dec 06 03:36:45 crc kubenswrapper[4980]: I1206 03:36:45.482450 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vpb28\" (UniqueName: \"kubernetes.io/projected/56a102c7-f341-4076-89dc-44428e77c164-kube-api-access-vpb28\") pod \"redhat-operators-bl7hx\" (UID: \"56a102c7-f341-4076-89dc-44428e77c164\") " pod="openshift-marketplace/redhat-operators-bl7hx" Dec 06 03:36:45 crc kubenswrapper[4980]: I1206 03:36:45.482543 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/56a102c7-f341-4076-89dc-44428e77c164-utilities\") pod \"redhat-operators-bl7hx\" (UID: \"56a102c7-f341-4076-89dc-44428e77c164\") " pod="openshift-marketplace/redhat-operators-bl7hx" Dec 06 03:36:45 crc kubenswrapper[4980]: I1206 03:36:45.483286 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/56a102c7-f341-4076-89dc-44428e77c164-utilities\") pod \"redhat-operators-bl7hx\" (UID: \"56a102c7-f341-4076-89dc-44428e77c164\") " pod="openshift-marketplace/redhat-operators-bl7hx" Dec 06 03:36:45 crc kubenswrapper[4980]: I1206 03:36:45.484184 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/56a102c7-f341-4076-89dc-44428e77c164-catalog-content\") pod \"redhat-operators-bl7hx\" (UID: \"56a102c7-f341-4076-89dc-44428e77c164\") " pod="openshift-marketplace/redhat-operators-bl7hx" Dec 06 03:36:45 crc kubenswrapper[4980]: I1206 03:36:45.515732 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vpb28\" (UniqueName: \"kubernetes.io/projected/56a102c7-f341-4076-89dc-44428e77c164-kube-api-access-vpb28\") pod \"redhat-operators-bl7hx\" (UID: \"56a102c7-f341-4076-89dc-44428e77c164\") " pod="openshift-marketplace/redhat-operators-bl7hx" Dec 06 03:36:45 crc kubenswrapper[4980]: I1206 03:36:45.592029 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-bl7hx" Dec 06 03:36:45 crc kubenswrapper[4980]: I1206 03:36:45.631673 4980 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29416530-9hm9m" Dec 06 03:36:45 crc kubenswrapper[4980]: I1206 03:36:45.693523 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/03704eb9-227d-4985-87cf-acd66ae76037-secret-volume\") pod \"03704eb9-227d-4985-87cf-acd66ae76037\" (UID: \"03704eb9-227d-4985-87cf-acd66ae76037\") " Dec 06 03:36:45 crc kubenswrapper[4980]: I1206 03:36:45.693589 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2w6nc\" (UniqueName: \"kubernetes.io/projected/03704eb9-227d-4985-87cf-acd66ae76037-kube-api-access-2w6nc\") pod \"03704eb9-227d-4985-87cf-acd66ae76037\" (UID: \"03704eb9-227d-4985-87cf-acd66ae76037\") " Dec 06 03:36:45 crc kubenswrapper[4980]: I1206 03:36:45.693682 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/03704eb9-227d-4985-87cf-acd66ae76037-config-volume\") pod \"03704eb9-227d-4985-87cf-acd66ae76037\" (UID: \"03704eb9-227d-4985-87cf-acd66ae76037\") " Dec 06 03:36:45 crc kubenswrapper[4980]: I1206 03:36:45.694825 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/03704eb9-227d-4985-87cf-acd66ae76037-config-volume" (OuterVolumeSpecName: "config-volume") pod "03704eb9-227d-4985-87cf-acd66ae76037" (UID: "03704eb9-227d-4985-87cf-acd66ae76037"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 03:36:45 crc kubenswrapper[4980]: I1206 03:36:45.701888 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/03704eb9-227d-4985-87cf-acd66ae76037-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "03704eb9-227d-4985-87cf-acd66ae76037" (UID: "03704eb9-227d-4985-87cf-acd66ae76037"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 03:36:45 crc kubenswrapper[4980]: I1206 03:36:45.755085 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-qw2fb" Dec 06 03:36:45 crc kubenswrapper[4980]: I1206 03:36:45.776096 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/03704eb9-227d-4985-87cf-acd66ae76037-kube-api-access-2w6nc" (OuterVolumeSpecName: "kube-api-access-2w6nc") pod "03704eb9-227d-4985-87cf-acd66ae76037" (UID: "03704eb9-227d-4985-87cf-acd66ae76037"). InnerVolumeSpecName "kube-api-access-2w6nc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 03:36:45 crc kubenswrapper[4980]: I1206 03:36:45.795450 4980 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/03704eb9-227d-4985-87cf-acd66ae76037-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 06 03:36:45 crc kubenswrapper[4980]: I1206 03:36:45.795479 4980 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2w6nc\" (UniqueName: \"kubernetes.io/projected/03704eb9-227d-4985-87cf-acd66ae76037-kube-api-access-2w6nc\") on node \"crc\" DevicePath \"\"" Dec 06 03:36:45 crc kubenswrapper[4980]: I1206 03:36:45.795488 4980 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/03704eb9-227d-4985-87cf-acd66ae76037-config-volume\") on node \"crc\" DevicePath \"\"" Dec 06 03:36:45 crc kubenswrapper[4980]: I1206 03:36:45.954229 4980 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-m9vl6"] Dec 06 03:36:46 crc kubenswrapper[4980]: I1206 03:36:46.002677 4980 generic.go:334] "Generic (PLEG): container finished" podID="3caa51b4-a311-4498-ac9b-10c20989684a" containerID="863754232f8ecce26db582347b737e0fae6c66c705dcb1ed803f6da176d756e1" exitCode=0 Dec 06 03:36:46 crc kubenswrapper[4980]: I1206 03:36:46.002769 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xhnw8" event={"ID":"3caa51b4-a311-4498-ac9b-10c20989684a","Type":"ContainerDied","Data":"863754232f8ecce26db582347b737e0fae6c66c705dcb1ed803f6da176d756e1"} Dec 06 03:36:46 crc kubenswrapper[4980]: I1206 03:36:46.002811 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xhnw8" event={"ID":"3caa51b4-a311-4498-ac9b-10c20989684a","Type":"ContainerStarted","Data":"db9e84b373a64698e756936369c3ff5ed646c18669d165181d4f42e254397e0d"} Dec 06 03:36:46 crc kubenswrapper[4980]: I1206 03:36:46.024998 4980 generic.go:334] "Generic (PLEG): container finished" podID="2ce9bc16-3d2a-4feb-82c1-0f42518196cf" containerID="c07bf9bd85f149a01c55d59978da7a3af66f9d797c3c9138c946857d0785f2ac" exitCode=0 Dec 06 03:36:46 crc kubenswrapper[4980]: I1206 03:36:46.025739 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-gxcqb" event={"ID":"2ce9bc16-3d2a-4feb-82c1-0f42518196cf","Type":"ContainerDied","Data":"c07bf9bd85f149a01c55d59978da7a3af66f9d797c3c9138c946857d0785f2ac"} Dec 06 03:36:46 crc kubenswrapper[4980]: I1206 03:36:46.025768 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-gxcqb" event={"ID":"2ce9bc16-3d2a-4feb-82c1-0f42518196cf","Type":"ContainerStarted","Data":"88edfcbb22a3aa98874380591fdb3d5827b4fb84a7c528e82e509639ce7804f8"} Dec 06 03:36:46 crc kubenswrapper[4980]: W1206 03:36:46.039973 4980 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod43112506_9fdd_4b19_a5a5_caf864c09774.slice/crio-29df7b7fb5fdf33ae3bf5ebbe15b646210dafb41c0c00f8e8ab3338bc1294d88 WatchSource:0}: Error finding container 29df7b7fb5fdf33ae3bf5ebbe15b646210dafb41c0c00f8e8ab3338bc1294d88: Status 404 returned error can't find the container with id 29df7b7fb5fdf33ae3bf5ebbe15b646210dafb41c0c00f8e8ab3338bc1294d88 Dec 06 03:36:46 crc kubenswrapper[4980]: I1206 03:36:46.040601 4980 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29416530-9hm9m" Dec 06 03:36:46 crc kubenswrapper[4980]: I1206 03:36:46.040783 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29416530-9hm9m" event={"ID":"03704eb9-227d-4985-87cf-acd66ae76037","Type":"ContainerDied","Data":"bf422279b8c14c2fa296df47fdefd0d0a736c7eb49fc94d969d975ec797735c0"} Dec 06 03:36:46 crc kubenswrapper[4980]: I1206 03:36:46.040823 4980 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="bf422279b8c14c2fa296df47fdefd0d0a736c7eb49fc94d969d975ec797735c0" Dec 06 03:36:46 crc kubenswrapper[4980]: I1206 03:36:46.084302 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-74hx2" event={"ID":"9b9bddd5-f192-4b4c-81ee-b523bf9c8705","Type":"ContainerStarted","Data":"dbcdb0ca2f7c0170ca4905dade6a542a19e8923a78614f7c0cff9cdd379e2119"} Dec 06 03:36:46 crc kubenswrapper[4980]: I1206 03:36:46.195631 4980 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="hostpath-provisioner/csi-hostpathplugin-74hx2" podStartSLOduration=15.195581201 podStartE2EDuration="15.195581201s" podCreationTimestamp="2025-12-06 03:36:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 03:36:46.104548013 +0000 UTC m=+165.349299294" watchObservedRunningTime="2025-12-06 03:36:46.195581201 +0000 UTC m=+165.440332492" Dec 06 03:36:46 crc kubenswrapper[4980]: I1206 03:36:46.202685 4980 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Dec 06 03:36:46 crc kubenswrapper[4980]: E1206 03:36:46.203059 4980 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="03704eb9-227d-4985-87cf-acd66ae76037" containerName="collect-profiles" Dec 06 03:36:46 crc kubenswrapper[4980]: I1206 03:36:46.203080 4980 state_mem.go:107] "Deleted CPUSet assignment" podUID="03704eb9-227d-4985-87cf-acd66ae76037" containerName="collect-profiles" Dec 06 03:36:46 crc kubenswrapper[4980]: I1206 03:36:46.203220 4980 memory_manager.go:354] "RemoveStaleState removing state" podUID="03704eb9-227d-4985-87cf-acd66ae76037" containerName="collect-profiles" Dec 06 03:36:46 crc kubenswrapper[4980]: I1206 03:36:46.208491 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 06 03:36:46 crc kubenswrapper[4980]: I1206 03:36:46.222341 4980 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager"/"kube-root-ca.crt" Dec 06 03:36:46 crc kubenswrapper[4980]: I1206 03:36:46.222916 4980 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager"/"installer-sa-dockercfg-kjl2n" Dec 06 03:36:46 crc kubenswrapper[4980]: I1206 03:36:46.228420 4980 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Dec 06 03:36:46 crc kubenswrapper[4980]: I1206 03:36:46.309480 4980 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-bl7hx"] Dec 06 03:36:46 crc kubenswrapper[4980]: I1206 03:36:46.377893 4980 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-qw2fb"] Dec 06 03:36:46 crc kubenswrapper[4980]: I1206 03:36:46.477369 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/ac8f73f9-1705-4c93-92f2-0f153c93d20e-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"ac8f73f9-1705-4c93-92f2-0f153c93d20e\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 06 03:36:46 crc kubenswrapper[4980]: I1206 03:36:46.477475 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/ac8f73f9-1705-4c93-92f2-0f153c93d20e-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"ac8f73f9-1705-4c93-92f2-0f153c93d20e\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 06 03:36:46 crc kubenswrapper[4980]: I1206 03:36:46.480573 4980 patch_prober.go:28] interesting pod/router-default-5444994796-pzzkd container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 06 03:36:46 crc kubenswrapper[4980]: [-]has-synced failed: reason withheld Dec 06 03:36:46 crc kubenswrapper[4980]: [+]process-running ok Dec 06 03:36:46 crc kubenswrapper[4980]: healthz check failed Dec 06 03:36:46 crc kubenswrapper[4980]: I1206 03:36:46.480649 4980 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-pzzkd" podUID="acb53ac6-3f09-424c-a830-a1e15eec0d62" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 06 03:36:46 crc kubenswrapper[4980]: W1206 03:36:46.483234 4980 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod56a102c7_f341_4076_89dc_44428e77c164.slice/crio-ed86090995deeef27f17707d16552c129a494499b8a80c3ec72ae12c80cc31b8 WatchSource:0}: Error finding container ed86090995deeef27f17707d16552c129a494499b8a80c3ec72ae12c80cc31b8: Status 404 returned error can't find the container with id ed86090995deeef27f17707d16552c129a494499b8a80c3ec72ae12c80cc31b8 Dec 06 03:36:46 crc kubenswrapper[4980]: I1206 03:36:46.596809 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/e6e88c5e-a255-449a-ae38-ac4d73a8e984-metrics-certs\") pod \"network-metrics-daemon-mcprh\" (UID: \"e6e88c5e-a255-449a-ae38-ac4d73a8e984\") " pod="openshift-multus/network-metrics-daemon-mcprh" Dec 06 03:36:46 crc kubenswrapper[4980]: I1206 03:36:46.597110 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/ac8f73f9-1705-4c93-92f2-0f153c93d20e-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"ac8f73f9-1705-4c93-92f2-0f153c93d20e\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 06 03:36:46 crc kubenswrapper[4980]: I1206 03:36:46.598210 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/ac8f73f9-1705-4c93-92f2-0f153c93d20e-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"ac8f73f9-1705-4c93-92f2-0f153c93d20e\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 06 03:36:46 crc kubenswrapper[4980]: I1206 03:36:46.598403 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/ac8f73f9-1705-4c93-92f2-0f153c93d20e-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"ac8f73f9-1705-4c93-92f2-0f153c93d20e\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 06 03:36:46 crc kubenswrapper[4980]: I1206 03:36:46.622878 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/e6e88c5e-a255-449a-ae38-ac4d73a8e984-metrics-certs\") pod \"network-metrics-daemon-mcprh\" (UID: \"e6e88c5e-a255-449a-ae38-ac4d73a8e984\") " pod="openshift-multus/network-metrics-daemon-mcprh" Dec 06 03:36:46 crc kubenswrapper[4980]: I1206 03:36:46.626673 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/ac8f73f9-1705-4c93-92f2-0f153c93d20e-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"ac8f73f9-1705-4c93-92f2-0f153c93d20e\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 06 03:36:46 crc kubenswrapper[4980]: I1206 03:36:46.855794 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 06 03:36:46 crc kubenswrapper[4980]: I1206 03:36:46.912832 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-mcprh" Dec 06 03:36:47 crc kubenswrapper[4980]: I1206 03:36:47.179124 4980 generic.go:334] "Generic (PLEG): container finished" podID="43112506-9fdd-4b19-a5a5-caf864c09774" containerID="97406efd5585dbfcd36479b31e92271c541ccac868fb34f708f0309bc89189c4" exitCode=0 Dec 06 03:36:47 crc kubenswrapper[4980]: I1206 03:36:47.179551 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-m9vl6" event={"ID":"43112506-9fdd-4b19-a5a5-caf864c09774","Type":"ContainerDied","Data":"97406efd5585dbfcd36479b31e92271c541ccac868fb34f708f0309bc89189c4"} Dec 06 03:36:47 crc kubenswrapper[4980]: I1206 03:36:47.179578 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-m9vl6" event={"ID":"43112506-9fdd-4b19-a5a5-caf864c09774","Type":"ContainerStarted","Data":"29df7b7fb5fdf33ae3bf5ebbe15b646210dafb41c0c00f8e8ab3338bc1294d88"} Dec 06 03:36:47 crc kubenswrapper[4980]: I1206 03:36:47.182789 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-qw2fb" event={"ID":"b25ba819-6e4e-4ca0-a570-c2726b8fbec5","Type":"ContainerStarted","Data":"6f1a12c733903ff72d2e17f9c24273a497ba9798f9450b0680a991fe69b18a14"} Dec 06 03:36:47 crc kubenswrapper[4980]: I1206 03:36:47.207713 4980 generic.go:334] "Generic (PLEG): container finished" podID="56a102c7-f341-4076-89dc-44428e77c164" containerID="10568cc585896cfd066308ee3d8cc363c77bac90cfad8e139177a4794fba172f" exitCode=0 Dec 06 03:36:47 crc kubenswrapper[4980]: I1206 03:36:47.207781 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bl7hx" event={"ID":"56a102c7-f341-4076-89dc-44428e77c164","Type":"ContainerDied","Data":"10568cc585896cfd066308ee3d8cc363c77bac90cfad8e139177a4794fba172f"} Dec 06 03:36:47 crc kubenswrapper[4980]: I1206 03:36:47.207829 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bl7hx" event={"ID":"56a102c7-f341-4076-89dc-44428e77c164","Type":"ContainerStarted","Data":"ed86090995deeef27f17707d16552c129a494499b8a80c3ec72ae12c80cc31b8"} Dec 06 03:36:48 crc kubenswrapper[4980]: I1206 03:36:48.153612 4980 patch_prober.go:28] interesting pod/router-default-5444994796-pzzkd container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 06 03:36:48 crc kubenswrapper[4980]: [-]has-synced failed: reason withheld Dec 06 03:36:48 crc kubenswrapper[4980]: [+]process-running ok Dec 06 03:36:48 crc kubenswrapper[4980]: healthz check failed Dec 06 03:36:48 crc kubenswrapper[4980]: I1206 03:36:48.153884 4980 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-pzzkd" podUID="acb53ac6-3f09-424c-a830-a1e15eec0d62" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 06 03:36:48 crc kubenswrapper[4980]: I1206 03:36:48.372470 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-qw2fb" event={"ID":"b25ba819-6e4e-4ca0-a570-c2726b8fbec5","Type":"ContainerStarted","Data":"9e89a295b67d24815fed2d32d770bfc36b4d92b99ab27d36045774ea8cf6e846"} Dec 06 03:36:48 crc kubenswrapper[4980]: I1206 03:36:48.423243 4980 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-697d97f7c8-qw2fb" podStartSLOduration=145.423186431 podStartE2EDuration="2m25.423186431s" podCreationTimestamp="2025-12-06 03:34:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 03:36:48.394903445 +0000 UTC m=+167.639654716" watchObservedRunningTime="2025-12-06 03:36:48.423186431 +0000 UTC m=+167.667937732" Dec 06 03:36:48 crc kubenswrapper[4980]: I1206 03:36:48.432198 4980 patch_prober.go:28] interesting pod/router-default-5444994796-pzzkd container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 06 03:36:48 crc kubenswrapper[4980]: [-]has-synced failed: reason withheld Dec 06 03:36:48 crc kubenswrapper[4980]: [+]process-running ok Dec 06 03:36:48 crc kubenswrapper[4980]: healthz check failed Dec 06 03:36:48 crc kubenswrapper[4980]: I1206 03:36:48.432317 4980 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-pzzkd" podUID="acb53ac6-3f09-424c-a830-a1e15eec0d62" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 06 03:36:48 crc kubenswrapper[4980]: I1206 03:36:48.665582 4980 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Dec 06 03:36:48 crc kubenswrapper[4980]: I1206 03:36:48.703842 4980 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-mcprh"] Dec 06 03:36:49 crc kubenswrapper[4980]: I1206 03:36:49.041667 4980 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-apiserver/apiserver-76f77b778f-s6h9g" Dec 06 03:36:49 crc kubenswrapper[4980]: I1206 03:36:49.045548 4980 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-apiserver/apiserver-76f77b778f-s6h9g" Dec 06 03:36:49 crc kubenswrapper[4980]: I1206 03:36:49.415189 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-mcprh" event={"ID":"e6e88c5e-a255-449a-ae38-ac4d73a8e984","Type":"ContainerStarted","Data":"aee1cfc8adb00151e386c2516ad77d83db7ac608567c1ec0a85065dbfb46e5cd"} Dec 06 03:36:49 crc kubenswrapper[4980]: I1206 03:36:49.426027 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"ac8f73f9-1705-4c93-92f2-0f153c93d20e","Type":"ContainerStarted","Data":"ed392f09b05ae7810c239fedea6e838d3cc7f93e865937ff138a203744904571"} Dec 06 03:36:49 crc kubenswrapper[4980]: I1206 03:36:49.426729 4980 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-697d97f7c8-qw2fb" Dec 06 03:36:49 crc kubenswrapper[4980]: I1206 03:36:49.428560 4980 patch_prober.go:28] interesting pod/router-default-5444994796-pzzkd container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 06 03:36:49 crc kubenswrapper[4980]: [-]has-synced failed: reason withheld Dec 06 03:36:49 crc kubenswrapper[4980]: [+]process-running ok Dec 06 03:36:49 crc kubenswrapper[4980]: healthz check failed Dec 06 03:36:49 crc kubenswrapper[4980]: I1206 03:36:49.428633 4980 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-pzzkd" podUID="acb53ac6-3f09-424c-a830-a1e15eec0d62" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 06 03:36:49 crc kubenswrapper[4980]: I1206 03:36:49.546356 4980 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Dec 06 03:36:49 crc kubenswrapper[4980]: I1206 03:36:49.547210 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 06 03:36:49 crc kubenswrapper[4980]: I1206 03:36:49.550045 4980 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Dec 06 03:36:49 crc kubenswrapper[4980]: I1206 03:36:49.551726 4980 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Dec 06 03:36:49 crc kubenswrapper[4980]: I1206 03:36:49.554362 4980 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Dec 06 03:36:49 crc kubenswrapper[4980]: I1206 03:36:49.732117 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/cb7c1fab-072e-49dc-a1d3-ccad60f758ed-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"cb7c1fab-072e-49dc-a1d3-ccad60f758ed\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 06 03:36:49 crc kubenswrapper[4980]: I1206 03:36:49.732218 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/cb7c1fab-072e-49dc-a1d3-ccad60f758ed-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"cb7c1fab-072e-49dc-a1d3-ccad60f758ed\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 06 03:36:49 crc kubenswrapper[4980]: I1206 03:36:49.833078 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/cb7c1fab-072e-49dc-a1d3-ccad60f758ed-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"cb7c1fab-072e-49dc-a1d3-ccad60f758ed\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 06 03:36:49 crc kubenswrapper[4980]: I1206 03:36:49.833220 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/cb7c1fab-072e-49dc-a1d3-ccad60f758ed-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"cb7c1fab-072e-49dc-a1d3-ccad60f758ed\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 06 03:36:49 crc kubenswrapper[4980]: I1206 03:36:49.833155 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/cb7c1fab-072e-49dc-a1d3-ccad60f758ed-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"cb7c1fab-072e-49dc-a1d3-ccad60f758ed\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 06 03:36:49 crc kubenswrapper[4980]: I1206 03:36:49.873936 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/cb7c1fab-072e-49dc-a1d3-ccad60f758ed-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"cb7c1fab-072e-49dc-a1d3-ccad60f758ed\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 06 03:36:49 crc kubenswrapper[4980]: I1206 03:36:49.885332 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 06 03:36:50 crc kubenswrapper[4980]: I1206 03:36:50.416463 4980 patch_prober.go:28] interesting pod/router-default-5444994796-pzzkd container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 06 03:36:50 crc kubenswrapper[4980]: [-]has-synced failed: reason withheld Dec 06 03:36:50 crc kubenswrapper[4980]: [+]process-running ok Dec 06 03:36:50 crc kubenswrapper[4980]: healthz check failed Dec 06 03:36:50 crc kubenswrapper[4980]: I1206 03:36:50.417030 4980 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-pzzkd" podUID="acb53ac6-3f09-424c-a830-a1e15eec0d62" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 06 03:36:50 crc kubenswrapper[4980]: I1206 03:36:50.476212 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-mcprh" event={"ID":"e6e88c5e-a255-449a-ae38-ac4d73a8e984","Type":"ContainerStarted","Data":"0fd0e45d00af55e0d9aea21327efcee3bfc9b6a0985c64a35e1d796d8860d1c0"} Dec 06 03:36:50 crc kubenswrapper[4980]: I1206 03:36:50.499939 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"ac8f73f9-1705-4c93-92f2-0f153c93d20e","Type":"ContainerStarted","Data":"c101088ee9fe367be384a1d6376199d7511d4bf74c383d442f98dd00c10cee64"} Dec 06 03:36:50 crc kubenswrapper[4980]: I1206 03:36:50.525632 4980 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/revision-pruner-9-crc" podStartSLOduration=4.5254627020000004 podStartE2EDuration="4.525462702s" podCreationTimestamp="2025-12-06 03:36:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 03:36:50.521374654 +0000 UTC m=+169.766125935" watchObservedRunningTime="2025-12-06 03:36:50.525462702 +0000 UTC m=+169.770213973" Dec 06 03:36:50 crc kubenswrapper[4980]: I1206 03:36:50.632012 4980 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Dec 06 03:36:51 crc kubenswrapper[4980]: I1206 03:36:51.527407 4980 patch_prober.go:28] interesting pod/router-default-5444994796-pzzkd container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 06 03:36:51 crc kubenswrapper[4980]: [-]has-synced failed: reason withheld Dec 06 03:36:51 crc kubenswrapper[4980]: [+]process-running ok Dec 06 03:36:51 crc kubenswrapper[4980]: healthz check failed Dec 06 03:36:51 crc kubenswrapper[4980]: I1206 03:36:51.527856 4980 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-pzzkd" podUID="acb53ac6-3f09-424c-a830-a1e15eec0d62" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 06 03:36:51 crc kubenswrapper[4980]: I1206 03:36:51.541495 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"cb7c1fab-072e-49dc-a1d3-ccad60f758ed","Type":"ContainerStarted","Data":"718bd119adec34a047a8a64d4a0812a8271e0f4970e589e97fe509167d481345"} Dec 06 03:36:51 crc kubenswrapper[4980]: I1206 03:36:51.543930 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-mcprh" event={"ID":"e6e88c5e-a255-449a-ae38-ac4d73a8e984","Type":"ContainerStarted","Data":"2235b30cc336cfde7308fec358ce17600e5d8f65fbcbc623e7e74e54d1b84d39"} Dec 06 03:36:51 crc kubenswrapper[4980]: I1206 03:36:51.546879 4980 generic.go:334] "Generic (PLEG): container finished" podID="ac8f73f9-1705-4c93-92f2-0f153c93d20e" containerID="c101088ee9fe367be384a1d6376199d7511d4bf74c383d442f98dd00c10cee64" exitCode=0 Dec 06 03:36:51 crc kubenswrapper[4980]: I1206 03:36:51.546930 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"ac8f73f9-1705-4c93-92f2-0f153c93d20e","Type":"ContainerDied","Data":"c101088ee9fe367be384a1d6376199d7511d4bf74c383d442f98dd00c10cee64"} Dec 06 03:36:51 crc kubenswrapper[4980]: I1206 03:36:51.559455 4980 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/network-metrics-daemon-mcprh" podStartSLOduration=149.55936244 podStartE2EDuration="2m29.55936244s" podCreationTimestamp="2025-12-06 03:34:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 03:36:51.556468457 +0000 UTC m=+170.801219738" watchObservedRunningTime="2025-12-06 03:36:51.55936244 +0000 UTC m=+170.804113721" Dec 06 03:36:52 crc kubenswrapper[4980]: I1206 03:36:52.415320 4980 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-ingress/router-default-5444994796-pzzkd" Dec 06 03:36:52 crc kubenswrapper[4980]: I1206 03:36:52.426725 4980 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ingress/router-default-5444994796-pzzkd" Dec 06 03:36:52 crc kubenswrapper[4980]: I1206 03:36:52.557680 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"cb7c1fab-072e-49dc-a1d3-ccad60f758ed","Type":"ContainerStarted","Data":"b331778037695b407859edf7fdfe573658df1df018eb31b16b8d38bcc43f291e"} Dec 06 03:36:52 crc kubenswrapper[4980]: I1206 03:36:52.583176 4980 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-dns/dns-default-4dcjh" Dec 06 03:36:52 crc kubenswrapper[4980]: I1206 03:36:52.605389 4980 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/revision-pruner-8-crc" podStartSLOduration=3.605367318 podStartE2EDuration="3.605367318s" podCreationTimestamp="2025-12-06 03:36:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 03:36:52.576818544 +0000 UTC m=+171.821569815" watchObservedRunningTime="2025-12-06 03:36:52.605367318 +0000 UTC m=+171.850118589" Dec 06 03:36:52 crc kubenswrapper[4980]: I1206 03:36:52.910789 4980 patch_prober.go:28] interesting pod/console-f9d7485db-gjmj8 container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.6:8443/health\": dial tcp 10.217.0.6:8443: connect: connection refused" start-of-body= Dec 06 03:36:52 crc kubenswrapper[4980]: I1206 03:36:52.910877 4980 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-gjmj8" podUID="443a2dc9-0546-4ab0-a448-eab81c09b482" containerName="console" probeResult="failure" output="Get \"https://10.217.0.6:8443/health\": dial tcp 10.217.0.6:8443: connect: connection refused" Dec 06 03:36:53 crc kubenswrapper[4980]: I1206 03:36:53.491197 4980 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 06 03:36:53 crc kubenswrapper[4980]: I1206 03:36:53.566795 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/ac8f73f9-1705-4c93-92f2-0f153c93d20e-kubelet-dir\") pod \"ac8f73f9-1705-4c93-92f2-0f153c93d20e\" (UID: \"ac8f73f9-1705-4c93-92f2-0f153c93d20e\") " Dec 06 03:36:53 crc kubenswrapper[4980]: I1206 03:36:53.566894 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/ac8f73f9-1705-4c93-92f2-0f153c93d20e-kube-api-access\") pod \"ac8f73f9-1705-4c93-92f2-0f153c93d20e\" (UID: \"ac8f73f9-1705-4c93-92f2-0f153c93d20e\") " Dec 06 03:36:53 crc kubenswrapper[4980]: I1206 03:36:53.568026 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ac8f73f9-1705-4c93-92f2-0f153c93d20e-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "ac8f73f9-1705-4c93-92f2-0f153c93d20e" (UID: "ac8f73f9-1705-4c93-92f2-0f153c93d20e"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 03:36:53 crc kubenswrapper[4980]: I1206 03:36:53.577888 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"ac8f73f9-1705-4c93-92f2-0f153c93d20e","Type":"ContainerDied","Data":"ed392f09b05ae7810c239fedea6e838d3cc7f93e865937ff138a203744904571"} Dec 06 03:36:53 crc kubenswrapper[4980]: I1206 03:36:53.578674 4980 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ed392f09b05ae7810c239fedea6e838d3cc7f93e865937ff138a203744904571" Dec 06 03:36:53 crc kubenswrapper[4980]: I1206 03:36:53.578749 4980 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 06 03:36:53 crc kubenswrapper[4980]: I1206 03:36:53.579820 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ac8f73f9-1705-4c93-92f2-0f153c93d20e-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "ac8f73f9-1705-4c93-92f2-0f153c93d20e" (UID: "ac8f73f9-1705-4c93-92f2-0f153c93d20e"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 03:36:53 crc kubenswrapper[4980]: I1206 03:36:53.676755 4980 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/ac8f73f9-1705-4c93-92f2-0f153c93d20e-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 06 03:36:53 crc kubenswrapper[4980]: I1206 03:36:53.676807 4980 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/ac8f73f9-1705-4c93-92f2-0f153c93d20e-kubelet-dir\") on node \"crc\" DevicePath \"\"" Dec 06 03:36:53 crc kubenswrapper[4980]: I1206 03:36:53.707632 4980 generic.go:334] "Generic (PLEG): container finished" podID="cb7c1fab-072e-49dc-a1d3-ccad60f758ed" containerID="b331778037695b407859edf7fdfe573658df1df018eb31b16b8d38bcc43f291e" exitCode=0 Dec 06 03:36:53 crc kubenswrapper[4980]: I1206 03:36:53.707684 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"cb7c1fab-072e-49dc-a1d3-ccad60f758ed","Type":"ContainerDied","Data":"b331778037695b407859edf7fdfe573658df1df018eb31b16b8d38bcc43f291e"} Dec 06 03:36:54 crc kubenswrapper[4980]: I1206 03:36:54.415476 4980 patch_prober.go:28] interesting pod/downloads-7954f5f757-6h96g container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.9:8080/\": dial tcp 10.217.0.9:8080: connect: connection refused" start-of-body= Dec 06 03:36:54 crc kubenswrapper[4980]: I1206 03:36:54.415590 4980 patch_prober.go:28] interesting pod/downloads-7954f5f757-6h96g container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.9:8080/\": dial tcp 10.217.0.9:8080: connect: connection refused" start-of-body= Dec 06 03:36:54 crc kubenswrapper[4980]: I1206 03:36:54.415595 4980 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-6h96g" podUID="346bb011-c34e-4286-90d3-0242205e8e94" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.9:8080/\": dial tcp 10.217.0.9:8080: connect: connection refused" Dec 06 03:36:54 crc kubenswrapper[4980]: I1206 03:36:54.415691 4980 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-6h96g" podUID="346bb011-c34e-4286-90d3-0242205e8e94" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.9:8080/\": dial tcp 10.217.0.9:8080: connect: connection refused" Dec 06 03:36:55 crc kubenswrapper[4980]: I1206 03:36:55.167150 4980 patch_prober.go:28] interesting pod/machine-config-daemon-r5zfb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 03:36:55 crc kubenswrapper[4980]: I1206 03:36:55.167653 4980 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-r5zfb" podUID="320f44d1-a671-4a91-b328-a8b0fdd8f23a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 03:37:00 crc kubenswrapper[4980]: I1206 03:37:00.646922 4980 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 03:37:02 crc kubenswrapper[4980]: I1206 03:37:02.909389 4980 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-f9d7485db-gjmj8" Dec 06 03:37:02 crc kubenswrapper[4980]: I1206 03:37:02.917000 4980 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-f9d7485db-gjmj8" Dec 06 03:37:04 crc kubenswrapper[4980]: I1206 03:37:04.298276 4980 patch_prober.go:28] interesting pod/downloads-7954f5f757-6h96g container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.9:8080/\": dial tcp 10.217.0.9:8080: connect: connection refused" start-of-body= Dec 06 03:37:04 crc kubenswrapper[4980]: I1206 03:37:04.298726 4980 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-6h96g" podUID="346bb011-c34e-4286-90d3-0242205e8e94" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.9:8080/\": dial tcp 10.217.0.9:8080: connect: connection refused" Dec 06 03:37:04 crc kubenswrapper[4980]: I1206 03:37:04.298788 4980 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-console/downloads-7954f5f757-6h96g" Dec 06 03:37:04 crc kubenswrapper[4980]: I1206 03:37:04.298652 4980 patch_prober.go:28] interesting pod/downloads-7954f5f757-6h96g container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.9:8080/\": dial tcp 10.217.0.9:8080: connect: connection refused" start-of-body= Dec 06 03:37:04 crc kubenswrapper[4980]: I1206 03:37:04.299055 4980 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-6h96g" podUID="346bb011-c34e-4286-90d3-0242205e8e94" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.9:8080/\": dial tcp 10.217.0.9:8080: connect: connection refused" Dec 06 03:37:04 crc kubenswrapper[4980]: I1206 03:37:04.299305 4980 patch_prober.go:28] interesting pod/downloads-7954f5f757-6h96g container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.9:8080/\": dial tcp 10.217.0.9:8080: connect: connection refused" start-of-body= Dec 06 03:37:04 crc kubenswrapper[4980]: I1206 03:37:04.299327 4980 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-6h96g" podUID="346bb011-c34e-4286-90d3-0242205e8e94" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.9:8080/\": dial tcp 10.217.0.9:8080: connect: connection refused" Dec 06 03:37:04 crc kubenswrapper[4980]: I1206 03:37:04.300870 4980 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="download-server" containerStatusID={"Type":"cri-o","ID":"c1bf646e1a1a4bf331bae0dde027acdbea896ad1d5a0d6c940f5a5a5d3155843"} pod="openshift-console/downloads-7954f5f757-6h96g" containerMessage="Container download-server failed liveness probe, will be restarted" Dec 06 03:37:04 crc kubenswrapper[4980]: I1206 03:37:04.301051 4980 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-console/downloads-7954f5f757-6h96g" podUID="346bb011-c34e-4286-90d3-0242205e8e94" containerName="download-server" containerID="cri-o://c1bf646e1a1a4bf331bae0dde027acdbea896ad1d5a0d6c940f5a5a5d3155843" gracePeriod=2 Dec 06 03:37:05 crc kubenswrapper[4980]: I1206 03:37:05.801080 4980 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-697d97f7c8-qw2fb" Dec 06 03:37:06 crc kubenswrapper[4980]: I1206 03:37:06.935089 4980 generic.go:334] "Generic (PLEG): container finished" podID="346bb011-c34e-4286-90d3-0242205e8e94" containerID="c1bf646e1a1a4bf331bae0dde027acdbea896ad1d5a0d6c940f5a5a5d3155843" exitCode=0 Dec 06 03:37:06 crc kubenswrapper[4980]: I1206 03:37:06.935194 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-6h96g" event={"ID":"346bb011-c34e-4286-90d3-0242205e8e94","Type":"ContainerDied","Data":"c1bf646e1a1a4bf331bae0dde027acdbea896ad1d5a0d6c940f5a5a5d3155843"} Dec 06 03:37:14 crc kubenswrapper[4980]: I1206 03:37:14.298320 4980 patch_prober.go:28] interesting pod/downloads-7954f5f757-6h96g container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.9:8080/\": dial tcp 10.217.0.9:8080: connect: connection refused" start-of-body= Dec 06 03:37:14 crc kubenswrapper[4980]: I1206 03:37:14.298692 4980 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-6h96g" podUID="346bb011-c34e-4286-90d3-0242205e8e94" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.9:8080/\": dial tcp 10.217.0.9:8080: connect: connection refused" Dec 06 03:37:14 crc kubenswrapper[4980]: I1206 03:37:14.803435 4980 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-qcn9f" Dec 06 03:37:17 crc kubenswrapper[4980]: I1206 03:37:17.135421 4980 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 06 03:37:17 crc kubenswrapper[4980]: I1206 03:37:17.235264 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/cb7c1fab-072e-49dc-a1d3-ccad60f758ed-kubelet-dir\") pod \"cb7c1fab-072e-49dc-a1d3-ccad60f758ed\" (UID: \"cb7c1fab-072e-49dc-a1d3-ccad60f758ed\") " Dec 06 03:37:17 crc kubenswrapper[4980]: I1206 03:37:17.235364 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/cb7c1fab-072e-49dc-a1d3-ccad60f758ed-kube-api-access\") pod \"cb7c1fab-072e-49dc-a1d3-ccad60f758ed\" (UID: \"cb7c1fab-072e-49dc-a1d3-ccad60f758ed\") " Dec 06 03:37:17 crc kubenswrapper[4980]: I1206 03:37:17.235530 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/cb7c1fab-072e-49dc-a1d3-ccad60f758ed-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "cb7c1fab-072e-49dc-a1d3-ccad60f758ed" (UID: "cb7c1fab-072e-49dc-a1d3-ccad60f758ed"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 03:37:17 crc kubenswrapper[4980]: I1206 03:37:17.235970 4980 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/cb7c1fab-072e-49dc-a1d3-ccad60f758ed-kubelet-dir\") on node \"crc\" DevicePath \"\"" Dec 06 03:37:17 crc kubenswrapper[4980]: I1206 03:37:17.255765 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cb7c1fab-072e-49dc-a1d3-ccad60f758ed-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "cb7c1fab-072e-49dc-a1d3-ccad60f758ed" (UID: "cb7c1fab-072e-49dc-a1d3-ccad60f758ed"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 03:37:17 crc kubenswrapper[4980]: I1206 03:37:17.337953 4980 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/cb7c1fab-072e-49dc-a1d3-ccad60f758ed-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 06 03:37:17 crc kubenswrapper[4980]: I1206 03:37:17.999401 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"cb7c1fab-072e-49dc-a1d3-ccad60f758ed","Type":"ContainerDied","Data":"718bd119adec34a047a8a64d4a0812a8271e0f4970e589e97fe509167d481345"} Dec 06 03:37:17 crc kubenswrapper[4980]: I1206 03:37:17.999453 4980 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="718bd119adec34a047a8a64d4a0812a8271e0f4970e589e97fe509167d481345" Dec 06 03:37:17 crc kubenswrapper[4980]: I1206 03:37:17.999538 4980 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 06 03:37:24 crc kubenswrapper[4980]: I1206 03:37:24.301267 4980 patch_prober.go:28] interesting pod/downloads-7954f5f757-6h96g container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.9:8080/\": dial tcp 10.217.0.9:8080: connect: connection refused" start-of-body= Dec 06 03:37:24 crc kubenswrapper[4980]: I1206 03:37:24.301855 4980 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-6h96g" podUID="346bb011-c34e-4286-90d3-0242205e8e94" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.9:8080/\": dial tcp 10.217.0.9:8080: connect: connection refused" Dec 06 03:37:24 crc kubenswrapper[4980]: I1206 03:37:24.318279 4980 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Dec 06 03:37:24 crc kubenswrapper[4980]: E1206 03:37:24.319274 4980 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ac8f73f9-1705-4c93-92f2-0f153c93d20e" containerName="pruner" Dec 06 03:37:24 crc kubenswrapper[4980]: I1206 03:37:24.319334 4980 state_mem.go:107] "Deleted CPUSet assignment" podUID="ac8f73f9-1705-4c93-92f2-0f153c93d20e" containerName="pruner" Dec 06 03:37:24 crc kubenswrapper[4980]: E1206 03:37:24.319372 4980 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cb7c1fab-072e-49dc-a1d3-ccad60f758ed" containerName="pruner" Dec 06 03:37:24 crc kubenswrapper[4980]: I1206 03:37:24.319384 4980 state_mem.go:107] "Deleted CPUSet assignment" podUID="cb7c1fab-072e-49dc-a1d3-ccad60f758ed" containerName="pruner" Dec 06 03:37:24 crc kubenswrapper[4980]: I1206 03:37:24.319585 4980 memory_manager.go:354] "RemoveStaleState removing state" podUID="cb7c1fab-072e-49dc-a1d3-ccad60f758ed" containerName="pruner" Dec 06 03:37:24 crc kubenswrapper[4980]: I1206 03:37:24.319605 4980 memory_manager.go:354] "RemoveStaleState removing state" podUID="ac8f73f9-1705-4c93-92f2-0f153c93d20e" containerName="pruner" Dec 06 03:37:24 crc kubenswrapper[4980]: I1206 03:37:24.321578 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 06 03:37:24 crc kubenswrapper[4980]: I1206 03:37:24.377381 4980 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Dec 06 03:37:24 crc kubenswrapper[4980]: I1206 03:37:24.377321 4980 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Dec 06 03:37:24 crc kubenswrapper[4980]: I1206 03:37:24.389033 4980 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Dec 06 03:37:24 crc kubenswrapper[4980]: I1206 03:37:24.596891 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/965405e7-b022-4b55-945e-3b4e7adbc5e5-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"965405e7-b022-4b55-945e-3b4e7adbc5e5\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 06 03:37:24 crc kubenswrapper[4980]: I1206 03:37:24.596964 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/965405e7-b022-4b55-945e-3b4e7adbc5e5-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"965405e7-b022-4b55-945e-3b4e7adbc5e5\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 06 03:37:24 crc kubenswrapper[4980]: I1206 03:37:24.698765 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/965405e7-b022-4b55-945e-3b4e7adbc5e5-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"965405e7-b022-4b55-945e-3b4e7adbc5e5\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 06 03:37:24 crc kubenswrapper[4980]: I1206 03:37:24.699015 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/965405e7-b022-4b55-945e-3b4e7adbc5e5-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"965405e7-b022-4b55-945e-3b4e7adbc5e5\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 06 03:37:24 crc kubenswrapper[4980]: I1206 03:37:24.699077 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/965405e7-b022-4b55-945e-3b4e7adbc5e5-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"965405e7-b022-4b55-945e-3b4e7adbc5e5\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 06 03:37:24 crc kubenswrapper[4980]: I1206 03:37:24.724685 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/965405e7-b022-4b55-945e-3b4e7adbc5e5-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"965405e7-b022-4b55-945e-3b4e7adbc5e5\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 06 03:37:25 crc kubenswrapper[4980]: I1206 03:37:25.001216 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 06 03:37:25 crc kubenswrapper[4980]: I1206 03:37:25.137572 4980 patch_prober.go:28] interesting pod/machine-config-daemon-r5zfb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 03:37:25 crc kubenswrapper[4980]: I1206 03:37:25.137681 4980 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-r5zfb" podUID="320f44d1-a671-4a91-b328-a8b0fdd8f23a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 03:37:25 crc kubenswrapper[4980]: I1206 03:37:25.137798 4980 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-r5zfb" Dec 06 03:37:25 crc kubenswrapper[4980]: I1206 03:37:25.138566 4980 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"52e8db644463794041978ced2c0591c13e0ada8e5f1a78a94a0a665218a37b5f"} pod="openshift-machine-config-operator/machine-config-daemon-r5zfb" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 06 03:37:25 crc kubenswrapper[4980]: I1206 03:37:25.138677 4980 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-r5zfb" podUID="320f44d1-a671-4a91-b328-a8b0fdd8f23a" containerName="machine-config-daemon" containerID="cri-o://52e8db644463794041978ced2c0591c13e0ada8e5f1a78a94a0a665218a37b5f" gracePeriod=600 Dec 06 03:37:25 crc kubenswrapper[4980]: I1206 03:37:25.287805 4980 patch_prober.go:28] interesting pod/authentication-operator-69f744f599-84ft2 container/authentication-operator namespace/openshift-authentication-operator: Liveness probe status=failure output="Get \"https://10.217.0.19:8443/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 06 03:37:25 crc kubenswrapper[4980]: I1206 03:37:25.287892 4980 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-authentication-operator/authentication-operator-69f744f599-84ft2" podUID="a3673c8a-c303-42ac-8565-f03e5d457aa1" containerName="authentication-operator" probeResult="failure" output="Get \"https://10.217.0.19:8443/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 06 03:37:27 crc kubenswrapper[4980]: I1206 03:37:27.102418 4980 generic.go:334] "Generic (PLEG): container finished" podID="320f44d1-a671-4a91-b328-a8b0fdd8f23a" containerID="52e8db644463794041978ced2c0591c13e0ada8e5f1a78a94a0a665218a37b5f" exitCode=0 Dec 06 03:37:27 crc kubenswrapper[4980]: I1206 03:37:27.102563 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-r5zfb" event={"ID":"320f44d1-a671-4a91-b328-a8b0fdd8f23a","Type":"ContainerDied","Data":"52e8db644463794041978ced2c0591c13e0ada8e5f1a78a94a0a665218a37b5f"} Dec 06 03:37:29 crc kubenswrapper[4980]: I1206 03:37:29.311501 4980 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Dec 06 03:37:29 crc kubenswrapper[4980]: I1206 03:37:29.313263 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Dec 06 03:37:29 crc kubenswrapper[4980]: I1206 03:37:29.328242 4980 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Dec 06 03:37:29 crc kubenswrapper[4980]: I1206 03:37:29.460841 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/dee63a36-cd1b-410c-b329-bd7966b1d38d-kubelet-dir\") pod \"installer-9-crc\" (UID: \"dee63a36-cd1b-410c-b329-bd7966b1d38d\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 06 03:37:29 crc kubenswrapper[4980]: I1206 03:37:29.460937 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/dee63a36-cd1b-410c-b329-bd7966b1d38d-kube-api-access\") pod \"installer-9-crc\" (UID: \"dee63a36-cd1b-410c-b329-bd7966b1d38d\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 06 03:37:29 crc kubenswrapper[4980]: I1206 03:37:29.461380 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/dee63a36-cd1b-410c-b329-bd7966b1d38d-var-lock\") pod \"installer-9-crc\" (UID: \"dee63a36-cd1b-410c-b329-bd7966b1d38d\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 06 03:37:29 crc kubenswrapper[4980]: I1206 03:37:29.562727 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/dee63a36-cd1b-410c-b329-bd7966b1d38d-var-lock\") pod \"installer-9-crc\" (UID: \"dee63a36-cd1b-410c-b329-bd7966b1d38d\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 06 03:37:29 crc kubenswrapper[4980]: I1206 03:37:29.562819 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/dee63a36-cd1b-410c-b329-bd7966b1d38d-kubelet-dir\") pod \"installer-9-crc\" (UID: \"dee63a36-cd1b-410c-b329-bd7966b1d38d\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 06 03:37:29 crc kubenswrapper[4980]: I1206 03:37:29.562900 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/dee63a36-cd1b-410c-b329-bd7966b1d38d-kube-api-access\") pod \"installer-9-crc\" (UID: \"dee63a36-cd1b-410c-b329-bd7966b1d38d\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 06 03:37:29 crc kubenswrapper[4980]: I1206 03:37:29.562982 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/dee63a36-cd1b-410c-b329-bd7966b1d38d-var-lock\") pod \"installer-9-crc\" (UID: \"dee63a36-cd1b-410c-b329-bd7966b1d38d\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 06 03:37:29 crc kubenswrapper[4980]: I1206 03:37:29.563106 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/dee63a36-cd1b-410c-b329-bd7966b1d38d-kubelet-dir\") pod \"installer-9-crc\" (UID: \"dee63a36-cd1b-410c-b329-bd7966b1d38d\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 06 03:37:29 crc kubenswrapper[4980]: I1206 03:37:29.583211 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/dee63a36-cd1b-410c-b329-bd7966b1d38d-kube-api-access\") pod \"installer-9-crc\" (UID: \"dee63a36-cd1b-410c-b329-bd7966b1d38d\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 06 03:37:29 crc kubenswrapper[4980]: I1206 03:37:29.654163 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Dec 06 03:37:34 crc kubenswrapper[4980]: I1206 03:37:34.342803 4980 patch_prober.go:28] interesting pod/downloads-7954f5f757-6h96g container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.9:8080/\": dial tcp 10.217.0.9:8080: connect: connection refused" start-of-body= Dec 06 03:37:34 crc kubenswrapper[4980]: I1206 03:37:34.343168 4980 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-6h96g" podUID="346bb011-c34e-4286-90d3-0242205e8e94" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.9:8080/\": dial tcp 10.217.0.9:8080: connect: connection refused" Dec 06 03:37:37 crc kubenswrapper[4980]: E1206 03:37:37.815484 4980 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Dec 06 03:37:37 crc kubenswrapper[4980]: E1206 03:37:37.816036 4980 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-vpb28,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-bl7hx_openshift-marketplace(56a102c7-f341-4076-89dc-44428e77c164): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 06 03:37:37 crc kubenswrapper[4980]: E1206 03:37:37.817324 4980 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-operators-bl7hx" podUID="56a102c7-f341-4076-89dc-44428e77c164" Dec 06 03:37:39 crc kubenswrapper[4980]: E1206 03:37:39.971986 4980 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-bl7hx" podUID="56a102c7-f341-4076-89dc-44428e77c164" Dec 06 03:37:40 crc kubenswrapper[4980]: E1206 03:37:40.868339 4980 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Dec 06 03:37:40 crc kubenswrapper[4980]: E1206 03:37:40.869026 4980 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-cnnh8,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-kp68b_openshift-marketplace(769abfb5-c18a-4cf0-a033-95b3a808867b): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 06 03:37:40 crc kubenswrapper[4980]: E1206 03:37:40.870286 4980 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-kp68b" podUID="769abfb5-c18a-4cf0-a033-95b3a808867b" Dec 06 03:37:43 crc kubenswrapper[4980]: E1206 03:37:43.072766 4980 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-kp68b" podUID="769abfb5-c18a-4cf0-a033-95b3a808867b" Dec 06 03:37:43 crc kubenswrapper[4980]: E1206 03:37:43.163026 4980 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Dec 06 03:37:43 crc kubenswrapper[4980]: E1206 03:37:43.163207 4980 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-zgc7q,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-xhnw8_openshift-marketplace(3caa51b4-a311-4498-ac9b-10c20989684a): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 06 03:37:43 crc kubenswrapper[4980]: E1206 03:37:43.165136 4980 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-xhnw8" podUID="3caa51b4-a311-4498-ac9b-10c20989684a" Dec 06 03:37:43 crc kubenswrapper[4980]: E1206 03:37:43.206035 4980 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Dec 06 03:37:43 crc kubenswrapper[4980]: E1206 03:37:43.206199 4980 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-4qvrf,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-58d94_openshift-marketplace(e965bb7a-1bf0-4574-aae4-2a45c75fb11c): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 06 03:37:43 crc kubenswrapper[4980]: E1206 03:37:43.207383 4980 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-58d94" podUID="e965bb7a-1bf0-4574-aae4-2a45c75fb11c" Dec 06 03:37:44 crc kubenswrapper[4980]: I1206 03:37:44.298250 4980 patch_prober.go:28] interesting pod/downloads-7954f5f757-6h96g container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.9:8080/\": dial tcp 10.217.0.9:8080: connect: connection refused" start-of-body= Dec 06 03:37:44 crc kubenswrapper[4980]: I1206 03:37:44.298859 4980 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-6h96g" podUID="346bb011-c34e-4286-90d3-0242205e8e94" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.9:8080/\": dial tcp 10.217.0.9:8080: connect: connection refused" Dec 06 03:37:44 crc kubenswrapper[4980]: E1206 03:37:44.584280 4980 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-xhnw8" podUID="3caa51b4-a311-4498-ac9b-10c20989684a" Dec 06 03:37:44 crc kubenswrapper[4980]: E1206 03:37:44.584892 4980 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-58d94" podUID="e965bb7a-1bf0-4574-aae4-2a45c75fb11c" Dec 06 03:37:44 crc kubenswrapper[4980]: E1206 03:37:44.682820 4980 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Dec 06 03:37:44 crc kubenswrapper[4980]: E1206 03:37:44.683310 4980 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-ppcll,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-mkd2m_openshift-marketplace(9c8b3374-77eb-44ad-b270-2263c6c23e28): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 06 03:37:44 crc kubenswrapper[4980]: E1206 03:37:44.684826 4980 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-mkd2m" podUID="9c8b3374-77eb-44ad-b270-2263c6c23e28" Dec 06 03:37:44 crc kubenswrapper[4980]: E1206 03:37:44.705303 4980 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Dec 06 03:37:44 crc kubenswrapper[4980]: E1206 03:37:44.705454 4980 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-hwq2v,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-m9vl6_openshift-marketplace(43112506-9fdd-4b19-a5a5-caf864c09774): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 06 03:37:44 crc kubenswrapper[4980]: E1206 03:37:44.706660 4980 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-operators-m9vl6" podUID="43112506-9fdd-4b19-a5a5-caf864c09774" Dec 06 03:37:44 crc kubenswrapper[4980]: E1206 03:37:44.755814 4980 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Dec 06 03:37:44 crc kubenswrapper[4980]: E1206 03:37:44.755968 4980 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-wxr7g,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-2vfzc_openshift-marketplace(652a7feb-30e3-4e9f-b1a9-c8ef201c658c): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 06 03:37:44 crc kubenswrapper[4980]: E1206 03:37:44.757538 4980 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-2vfzc" podUID="652a7feb-30e3-4e9f-b1a9-c8ef201c658c" Dec 06 03:37:44 crc kubenswrapper[4980]: E1206 03:37:44.852826 4980 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Dec 06 03:37:44 crc kubenswrapper[4980]: E1206 03:37:44.853290 4980 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-b9kxc,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-gxcqb_openshift-marketplace(2ce9bc16-3d2a-4feb-82c1-0f42518196cf): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 06 03:37:44 crc kubenswrapper[4980]: E1206 03:37:44.854485 4980 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-gxcqb" podUID="2ce9bc16-3d2a-4feb-82c1-0f42518196cf" Dec 06 03:37:45 crc kubenswrapper[4980]: I1206 03:37:45.089701 4980 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Dec 06 03:37:45 crc kubenswrapper[4980]: I1206 03:37:45.174073 4980 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Dec 06 03:37:45 crc kubenswrapper[4980]: I1206 03:37:45.221680 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"965405e7-b022-4b55-945e-3b4e7adbc5e5","Type":"ContainerStarted","Data":"530b2e47ddf20f16dc1dd84eac0eda386299e1d6a0f54719f5a6ca8361ee9ffc"} Dec 06 03:37:45 crc kubenswrapper[4980]: I1206 03:37:45.223766 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-6h96g" event={"ID":"346bb011-c34e-4286-90d3-0242205e8e94","Type":"ContainerStarted","Data":"d245e738e9bb6757d5834a4a3d90066ce5d403fe70e86aa8a5dd10634378af19"} Dec 06 03:37:45 crc kubenswrapper[4980]: I1206 03:37:45.224336 4980 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/downloads-7954f5f757-6h96g" Dec 06 03:37:45 crc kubenswrapper[4980]: I1206 03:37:45.224656 4980 patch_prober.go:28] interesting pod/downloads-7954f5f757-6h96g container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.9:8080/\": dial tcp 10.217.0.9:8080: connect: connection refused" start-of-body= Dec 06 03:37:45 crc kubenswrapper[4980]: I1206 03:37:45.224819 4980 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-6h96g" podUID="346bb011-c34e-4286-90d3-0242205e8e94" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.9:8080/\": dial tcp 10.217.0.9:8080: connect: connection refused" Dec 06 03:37:45 crc kubenswrapper[4980]: I1206 03:37:45.230350 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"dee63a36-cd1b-410c-b329-bd7966b1d38d","Type":"ContainerStarted","Data":"1bb044f31576bca8e79d7140ce8c441266db23773b959fe99d8cc5a6e40f2462"} Dec 06 03:37:45 crc kubenswrapper[4980]: I1206 03:37:45.250606 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-r5zfb" event={"ID":"320f44d1-a671-4a91-b328-a8b0fdd8f23a","Type":"ContainerStarted","Data":"fecbfcc218b13cf32c394772ff9a934bdde92dcfdcdb184849fd10fe2e85db2e"} Dec 06 03:37:45 crc kubenswrapper[4980]: E1206 03:37:45.254173 4980 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-m9vl6" podUID="43112506-9fdd-4b19-a5a5-caf864c09774" Dec 06 03:37:45 crc kubenswrapper[4980]: E1206 03:37:45.254437 4980 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-gxcqb" podUID="2ce9bc16-3d2a-4feb-82c1-0f42518196cf" Dec 06 03:37:45 crc kubenswrapper[4980]: E1206 03:37:45.254570 4980 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-mkd2m" podUID="9c8b3374-77eb-44ad-b270-2263c6c23e28" Dec 06 03:37:45 crc kubenswrapper[4980]: E1206 03:37:45.257718 4980 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-2vfzc" podUID="652a7feb-30e3-4e9f-b1a9-c8ef201c658c" Dec 06 03:37:46 crc kubenswrapper[4980]: I1206 03:37:46.257569 4980 patch_prober.go:28] interesting pod/downloads-7954f5f757-6h96g container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.9:8080/\": dial tcp 10.217.0.9:8080: connect: connection refused" start-of-body= Dec 06 03:37:46 crc kubenswrapper[4980]: I1206 03:37:46.260645 4980 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-6h96g" podUID="346bb011-c34e-4286-90d3-0242205e8e94" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.9:8080/\": dial tcp 10.217.0.9:8080: connect: connection refused" Dec 06 03:37:47 crc kubenswrapper[4980]: I1206 03:37:47.262210 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"965405e7-b022-4b55-945e-3b4e7adbc5e5","Type":"ContainerStarted","Data":"e68ee7a7f366ad27921593fbed9b4b770cd778675d62a6ce7a40b39eb70d4873"} Dec 06 03:37:48 crc kubenswrapper[4980]: I1206 03:37:48.269892 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"dee63a36-cd1b-410c-b329-bd7966b1d38d","Type":"ContainerStarted","Data":"61478a7d59badbae7bb924abe9d3ca7dd0623837917b65a27b2d83d7521a73cb"} Dec 06 03:37:49 crc kubenswrapper[4980]: I1206 03:37:49.275665 4980 generic.go:334] "Generic (PLEG): container finished" podID="965405e7-b022-4b55-945e-3b4e7adbc5e5" containerID="e68ee7a7f366ad27921593fbed9b4b770cd778675d62a6ce7a40b39eb70d4873" exitCode=0 Dec 06 03:37:49 crc kubenswrapper[4980]: I1206 03:37:49.275854 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"965405e7-b022-4b55-945e-3b4e7adbc5e5","Type":"ContainerDied","Data":"e68ee7a7f366ad27921593fbed9b4b770cd778675d62a6ce7a40b39eb70d4873"} Dec 06 03:37:49 crc kubenswrapper[4980]: I1206 03:37:49.295734 4980 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/revision-pruner-9-crc" podStartSLOduration=25.295692971 podStartE2EDuration="25.295692971s" podCreationTimestamp="2025-12-06 03:37:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 03:37:48.291985334 +0000 UTC m=+227.536736605" watchObservedRunningTime="2025-12-06 03:37:49.295692971 +0000 UTC m=+228.540444242" Dec 06 03:37:49 crc kubenswrapper[4980]: I1206 03:37:49.298498 4980 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/installer-9-crc" podStartSLOduration=20.298483872 podStartE2EDuration="20.298483872s" podCreationTimestamp="2025-12-06 03:37:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 03:37:49.293551778 +0000 UTC m=+228.538303069" watchObservedRunningTime="2025-12-06 03:37:49.298483872 +0000 UTC m=+228.543235143" Dec 06 03:37:50 crc kubenswrapper[4980]: I1206 03:37:50.633273 4980 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 06 03:37:50 crc kubenswrapper[4980]: I1206 03:37:50.750134 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/965405e7-b022-4b55-945e-3b4e7adbc5e5-kubelet-dir\") pod \"965405e7-b022-4b55-945e-3b4e7adbc5e5\" (UID: \"965405e7-b022-4b55-945e-3b4e7adbc5e5\") " Dec 06 03:37:50 crc kubenswrapper[4980]: I1206 03:37:50.750215 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/965405e7-b022-4b55-945e-3b4e7adbc5e5-kube-api-access\") pod \"965405e7-b022-4b55-945e-3b4e7adbc5e5\" (UID: \"965405e7-b022-4b55-945e-3b4e7adbc5e5\") " Dec 06 03:37:50 crc kubenswrapper[4980]: I1206 03:37:50.750250 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/965405e7-b022-4b55-945e-3b4e7adbc5e5-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "965405e7-b022-4b55-945e-3b4e7adbc5e5" (UID: "965405e7-b022-4b55-945e-3b4e7adbc5e5"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 03:37:50 crc kubenswrapper[4980]: I1206 03:37:50.750460 4980 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/965405e7-b022-4b55-945e-3b4e7adbc5e5-kubelet-dir\") on node \"crc\" DevicePath \"\"" Dec 06 03:37:50 crc kubenswrapper[4980]: I1206 03:37:50.757678 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/965405e7-b022-4b55-945e-3b4e7adbc5e5-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "965405e7-b022-4b55-945e-3b4e7adbc5e5" (UID: "965405e7-b022-4b55-945e-3b4e7adbc5e5"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 03:37:50 crc kubenswrapper[4980]: I1206 03:37:50.851546 4980 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/965405e7-b022-4b55-945e-3b4e7adbc5e5-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 06 03:37:51 crc kubenswrapper[4980]: I1206 03:37:51.287419 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"965405e7-b022-4b55-945e-3b4e7adbc5e5","Type":"ContainerDied","Data":"530b2e47ddf20f16dc1dd84eac0eda386299e1d6a0f54719f5a6ca8361ee9ffc"} Dec 06 03:37:51 crc kubenswrapper[4980]: I1206 03:37:51.287976 4980 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="530b2e47ddf20f16dc1dd84eac0eda386299e1d6a0f54719f5a6ca8361ee9ffc" Dec 06 03:37:51 crc kubenswrapper[4980]: I1206 03:37:51.287486 4980 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 06 03:37:54 crc kubenswrapper[4980]: I1206 03:37:54.303207 4980 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/downloads-7954f5f757-6h96g" Dec 06 03:37:58 crc kubenswrapper[4980]: I1206 03:37:58.332457 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bl7hx" event={"ID":"56a102c7-f341-4076-89dc-44428e77c164","Type":"ContainerStarted","Data":"6226b5de172f7a25afe31ae6e38b1c241129340606a18fafb76921d2e95f2958"} Dec 06 03:37:59 crc kubenswrapper[4980]: I1206 03:37:59.467948 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-58d94" event={"ID":"e965bb7a-1bf0-4574-aae4-2a45c75fb11c","Type":"ContainerStarted","Data":"8aa56292c24f69d83a0939e746f3716f1409df631c637167480a49c129beb2ca"} Dec 06 03:37:59 crc kubenswrapper[4980]: I1206 03:37:59.470935 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xhnw8" event={"ID":"3caa51b4-a311-4498-ac9b-10c20989684a","Type":"ContainerStarted","Data":"3fd75664d84b78f5765b694a48cb11d62122b84f14e3906a4eafb96f5bd2011c"} Dec 06 03:37:59 crc kubenswrapper[4980]: I1206 03:37:59.478963 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mkd2m" event={"ID":"9c8b3374-77eb-44ad-b270-2263c6c23e28","Type":"ContainerStarted","Data":"402a4b6521a47e3cfb5a8ca7a94ee76b83c72ac0e2c6645c91e7fcd6bee0b4bc"} Dec 06 03:37:59 crc kubenswrapper[4980]: I1206 03:37:59.484376 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-gxcqb" event={"ID":"2ce9bc16-3d2a-4feb-82c1-0f42518196cf","Type":"ContainerStarted","Data":"1dacbca162fc3e3abb85ba1b986d4c3a0c7e310dc145b1a6f971a248a74905e8"} Dec 06 03:37:59 crc kubenswrapper[4980]: I1206 03:37:59.485691 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kp68b" event={"ID":"769abfb5-c18a-4cf0-a033-95b3a808867b","Type":"ContainerStarted","Data":"db54300afb9839558c5c8d46ba1ba9aa3fac94f6b6e71f3c6511f74fd44c97ad"} Dec 06 03:38:00 crc kubenswrapper[4980]: I1206 03:38:00.575643 4980 generic.go:334] "Generic (PLEG): container finished" podID="3caa51b4-a311-4498-ac9b-10c20989684a" containerID="3fd75664d84b78f5765b694a48cb11d62122b84f14e3906a4eafb96f5bd2011c" exitCode=0 Dec 06 03:38:00 crc kubenswrapper[4980]: I1206 03:38:00.575688 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xhnw8" event={"ID":"3caa51b4-a311-4498-ac9b-10c20989684a","Type":"ContainerDied","Data":"3fd75664d84b78f5765b694a48cb11d62122b84f14e3906a4eafb96f5bd2011c"} Dec 06 03:38:00 crc kubenswrapper[4980]: I1206 03:38:00.577703 4980 generic.go:334] "Generic (PLEG): container finished" podID="2ce9bc16-3d2a-4feb-82c1-0f42518196cf" containerID="1dacbca162fc3e3abb85ba1b986d4c3a0c7e310dc145b1a6f971a248a74905e8" exitCode=0 Dec 06 03:38:00 crc kubenswrapper[4980]: I1206 03:38:00.577798 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-gxcqb" event={"ID":"2ce9bc16-3d2a-4feb-82c1-0f42518196cf","Type":"ContainerDied","Data":"1dacbca162fc3e3abb85ba1b986d4c3a0c7e310dc145b1a6f971a248a74905e8"} Dec 06 03:38:01 crc kubenswrapper[4980]: I1206 03:38:01.593591 4980 generic.go:334] "Generic (PLEG): container finished" podID="9c8b3374-77eb-44ad-b270-2263c6c23e28" containerID="402a4b6521a47e3cfb5a8ca7a94ee76b83c72ac0e2c6645c91e7fcd6bee0b4bc" exitCode=0 Dec 06 03:38:01 crc kubenswrapper[4980]: I1206 03:38:01.594306 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mkd2m" event={"ID":"9c8b3374-77eb-44ad-b270-2263c6c23e28","Type":"ContainerDied","Data":"402a4b6521a47e3cfb5a8ca7a94ee76b83c72ac0e2c6645c91e7fcd6bee0b4bc"} Dec 06 03:38:01 crc kubenswrapper[4980]: I1206 03:38:01.596291 4980 generic.go:334] "Generic (PLEG): container finished" podID="56a102c7-f341-4076-89dc-44428e77c164" containerID="6226b5de172f7a25afe31ae6e38b1c241129340606a18fafb76921d2e95f2958" exitCode=0 Dec 06 03:38:01 crc kubenswrapper[4980]: I1206 03:38:01.596349 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bl7hx" event={"ID":"56a102c7-f341-4076-89dc-44428e77c164","Type":"ContainerDied","Data":"6226b5de172f7a25afe31ae6e38b1c241129340606a18fafb76921d2e95f2958"} Dec 06 03:38:01 crc kubenswrapper[4980]: I1206 03:38:01.601184 4980 generic.go:334] "Generic (PLEG): container finished" podID="769abfb5-c18a-4cf0-a033-95b3a808867b" containerID="db54300afb9839558c5c8d46ba1ba9aa3fac94f6b6e71f3c6511f74fd44c97ad" exitCode=0 Dec 06 03:38:01 crc kubenswrapper[4980]: I1206 03:38:01.601256 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kp68b" event={"ID":"769abfb5-c18a-4cf0-a033-95b3a808867b","Type":"ContainerDied","Data":"db54300afb9839558c5c8d46ba1ba9aa3fac94f6b6e71f3c6511f74fd44c97ad"} Dec 06 03:38:01 crc kubenswrapper[4980]: I1206 03:38:01.606662 4980 generic.go:334] "Generic (PLEG): container finished" podID="e965bb7a-1bf0-4574-aae4-2a45c75fb11c" containerID="8aa56292c24f69d83a0939e746f3716f1409df631c637167480a49c129beb2ca" exitCode=0 Dec 06 03:38:01 crc kubenswrapper[4980]: I1206 03:38:01.606726 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-58d94" event={"ID":"e965bb7a-1bf0-4574-aae4-2a45c75fb11c","Type":"ContainerDied","Data":"8aa56292c24f69d83a0939e746f3716f1409df631c637167480a49c129beb2ca"} Dec 06 03:38:03 crc kubenswrapper[4980]: I1206 03:38:03.157996 4980 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-pmmcw"] Dec 06 03:38:24 crc kubenswrapper[4980]: I1206 03:38:24.733848 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-58d94" event={"ID":"e965bb7a-1bf0-4574-aae4-2a45c75fb11c","Type":"ContainerStarted","Data":"739e651d553aa530b056a94ecbc5131776fd3e6197e4a6840089a2c0f03e49b0"} Dec 06 03:38:24 crc kubenswrapper[4980]: I1206 03:38:24.739849 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2vfzc" event={"ID":"652a7feb-30e3-4e9f-b1a9-c8ef201c658c","Type":"ContainerStarted","Data":"ffbb4c8c6980bde657d7596d37a800b08dd5ac764045bebabcafe2c50bd30ac6"} Dec 06 03:38:24 crc kubenswrapper[4980]: I1206 03:38:24.742589 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xhnw8" event={"ID":"3caa51b4-a311-4498-ac9b-10c20989684a","Type":"ContainerStarted","Data":"e5e0adc7836a520047580410eb93beb81ac6a2dad1bc364b0c5bcf599f13cf86"} Dec 06 03:38:24 crc kubenswrapper[4980]: I1206 03:38:24.744928 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mkd2m" event={"ID":"9c8b3374-77eb-44ad-b270-2263c6c23e28","Type":"ContainerStarted","Data":"b3b94b4e2032df3e1ef26267e391f9b2dcba67ff9306a70e1944d4df7dc01e42"} Dec 06 03:38:24 crc kubenswrapper[4980]: I1206 03:38:24.747756 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bl7hx" event={"ID":"56a102c7-f341-4076-89dc-44428e77c164","Type":"ContainerStarted","Data":"fdff05d7789244fc432df351c27d2f5fcb050f2030386cc3a499b53e0c252358"} Dec 06 03:38:24 crc kubenswrapper[4980]: I1206 03:38:24.749994 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-gxcqb" event={"ID":"2ce9bc16-3d2a-4feb-82c1-0f42518196cf","Type":"ContainerStarted","Data":"9aa7d598bd03dc4b167375c94a61c6b2efcf6f1430a76effe832f91e012023d1"} Dec 06 03:38:24 crc kubenswrapper[4980]: I1206 03:38:24.752128 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kp68b" event={"ID":"769abfb5-c18a-4cf0-a033-95b3a808867b","Type":"ContainerStarted","Data":"d47b4f045fefdb2c7a2a6fe135c197172906e70ebfd6c2ae3c51b3798b9982cc"} Dec 06 03:38:24 crc kubenswrapper[4980]: I1206 03:38:24.753956 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-m9vl6" event={"ID":"43112506-9fdd-4b19-a5a5-caf864c09774","Type":"ContainerStarted","Data":"beb2e6aabb313ede1d7348363d791410e4e40f483c9f4bfceec700e83315d0b9"} Dec 06 03:38:24 crc kubenswrapper[4980]: I1206 03:38:24.768115 4980 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-58d94" podStartSLOduration=4.9992025380000005 podStartE2EDuration="1m43.768097725s" podCreationTimestamp="2025-12-06 03:36:41 +0000 UTC" firstStartedPulling="2025-12-06 03:36:44.872309249 +0000 UTC m=+164.117060520" lastFinishedPulling="2025-12-06 03:38:23.641204436 +0000 UTC m=+262.885955707" observedRunningTime="2025-12-06 03:38:24.766774227 +0000 UTC m=+264.011525498" watchObservedRunningTime="2025-12-06 03:38:24.768097725 +0000 UTC m=+264.012848996" Dec 06 03:38:24 crc kubenswrapper[4980]: I1206 03:38:24.844469 4980 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-gxcqb" podStartSLOduration=4.249211421 podStartE2EDuration="1m41.844445975s" podCreationTimestamp="2025-12-06 03:36:43 +0000 UTC" firstStartedPulling="2025-12-06 03:36:46.04107826 +0000 UTC m=+165.285829531" lastFinishedPulling="2025-12-06 03:38:23.636312804 +0000 UTC m=+262.881064085" observedRunningTime="2025-12-06 03:38:24.8414998 +0000 UTC m=+264.086251071" watchObservedRunningTime="2025-12-06 03:38:24.844445975 +0000 UTC m=+264.089197246" Dec 06 03:38:24 crc kubenswrapper[4980]: I1206 03:38:24.879070 4980 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-xhnw8" podStartSLOduration=3.308105709 podStartE2EDuration="1m40.879054107s" podCreationTimestamp="2025-12-06 03:36:44 +0000 UTC" firstStartedPulling="2025-12-06 03:36:46.009717895 +0000 UTC m=+165.254469166" lastFinishedPulling="2025-12-06 03:38:23.580666273 +0000 UTC m=+262.825417564" observedRunningTime="2025-12-06 03:38:24.874418263 +0000 UTC m=+264.119169534" watchObservedRunningTime="2025-12-06 03:38:24.879054107 +0000 UTC m=+264.123805378" Dec 06 03:38:24 crc kubenswrapper[4980]: I1206 03:38:24.918202 4980 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-mkd2m" podStartSLOduration=5.228984204 podStartE2EDuration="1m43.918181919s" podCreationTimestamp="2025-12-06 03:36:41 +0000 UTC" firstStartedPulling="2025-12-06 03:36:44.95164806 +0000 UTC m=+164.196399331" lastFinishedPulling="2025-12-06 03:38:23.640845775 +0000 UTC m=+262.885597046" observedRunningTime="2025-12-06 03:38:24.893648899 +0000 UTC m=+264.138400180" watchObservedRunningTime="2025-12-06 03:38:24.918181919 +0000 UTC m=+264.162933200" Dec 06 03:38:24 crc kubenswrapper[4980]: I1206 03:38:24.918967 4980 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-kp68b" podStartSLOduration=4.102969816 podStartE2EDuration="1m42.918960722s" podCreationTimestamp="2025-12-06 03:36:42 +0000 UTC" firstStartedPulling="2025-12-06 03:36:44.824119458 +0000 UTC m=+164.068870729" lastFinishedPulling="2025-12-06 03:38:23.640110364 +0000 UTC m=+262.884861635" observedRunningTime="2025-12-06 03:38:24.91508422 +0000 UTC m=+264.159835501" watchObservedRunningTime="2025-12-06 03:38:24.918960722 +0000 UTC m=+264.163711993" Dec 06 03:38:24 crc kubenswrapper[4980]: I1206 03:38:24.940438 4980 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-bl7hx" podStartSLOduration=3.601490806 podStartE2EDuration="1m39.940416223s" podCreationTimestamp="2025-12-06 03:36:45 +0000 UTC" firstStartedPulling="2025-12-06 03:36:47.208969658 +0000 UTC m=+166.453720929" lastFinishedPulling="2025-12-06 03:38:23.547895075 +0000 UTC m=+262.792646346" observedRunningTime="2025-12-06 03:38:24.938810267 +0000 UTC m=+264.183561558" watchObservedRunningTime="2025-12-06 03:38:24.940416223 +0000 UTC m=+264.185167494" Dec 06 03:38:25 crc kubenswrapper[4980]: I1206 03:38:25.543559 4980 kubelet.go:2431] "SyncLoop REMOVE" source="file" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Dec 06 03:38:25 crc kubenswrapper[4980]: I1206 03:38:25.544237 4980 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" containerID="cri-o://c28b6895b14e73f4f3eb7a8ce93076c05092ba9b58ea0e3e879c65bcfa227a21" gracePeriod=15 Dec 06 03:38:25 crc kubenswrapper[4980]: I1206 03:38:25.544316 4980 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" containerID="cri-o://76cff8a2d652df38b04eefc053832baf4cbb5db778ff5e525b18c2825141b118" gracePeriod=15 Dec 06 03:38:25 crc kubenswrapper[4980]: I1206 03:38:25.544338 4980 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" containerID="cri-o://d40e0e7446f8416bf9da94fc53c01be937f9a51a11eb101157d2ddd4597720d2" gracePeriod=15 Dec 06 03:38:25 crc kubenswrapper[4980]: I1206 03:38:25.544350 4980 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" containerID="cri-o://206f6f19fe3197f5d6798cc86c5473bf2decb02a2860ee7ffb5efb5f60952e8c" gracePeriod=15 Dec 06 03:38:25 crc kubenswrapper[4980]: I1206 03:38:25.544251 4980 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" containerID="cri-o://bafaa9670b05671fb57b658c2960bf3ee32c84e52f86a3a3bbab01d5a38ff0c4" gracePeriod=15 Dec 06 03:38:25 crc kubenswrapper[4980]: I1206 03:38:25.544825 4980 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Dec 06 03:38:25 crc kubenswrapper[4980]: E1206 03:38:25.546111 4980 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="setup" Dec 06 03:38:25 crc kubenswrapper[4980]: I1206 03:38:25.546215 4980 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="setup" Dec 06 03:38:25 crc kubenswrapper[4980]: E1206 03:38:25.546324 4980 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 06 03:38:25 crc kubenswrapper[4980]: I1206 03:38:25.546400 4980 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 06 03:38:25 crc kubenswrapper[4980]: E1206 03:38:25.546474 4980 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Dec 06 03:38:25 crc kubenswrapper[4980]: I1206 03:38:25.546579 4980 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Dec 06 03:38:25 crc kubenswrapper[4980]: E1206 03:38:25.546674 4980 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Dec 06 03:38:25 crc kubenswrapper[4980]: I1206 03:38:25.546826 4980 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Dec 06 03:38:25 crc kubenswrapper[4980]: E1206 03:38:25.546908 4980 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Dec 06 03:38:25 crc kubenswrapper[4980]: I1206 03:38:25.546982 4980 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Dec 06 03:38:25 crc kubenswrapper[4980]: E1206 03:38:25.547065 4980 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="965405e7-b022-4b55-945e-3b4e7adbc5e5" containerName="pruner" Dec 06 03:38:25 crc kubenswrapper[4980]: I1206 03:38:25.547157 4980 state_mem.go:107] "Deleted CPUSet assignment" podUID="965405e7-b022-4b55-945e-3b4e7adbc5e5" containerName="pruner" Dec 06 03:38:25 crc kubenswrapper[4980]: E1206 03:38:25.547232 4980 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 06 03:38:25 crc kubenswrapper[4980]: I1206 03:38:25.547308 4980 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 06 03:38:25 crc kubenswrapper[4980]: E1206 03:38:25.547392 4980 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Dec 06 03:38:25 crc kubenswrapper[4980]: I1206 03:38:25.547472 4980 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Dec 06 03:38:25 crc kubenswrapper[4980]: I1206 03:38:25.547694 4980 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Dec 06 03:38:25 crc kubenswrapper[4980]: I1206 03:38:25.547790 4980 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 06 03:38:25 crc kubenswrapper[4980]: I1206 03:38:25.547872 4980 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Dec 06 03:38:25 crc kubenswrapper[4980]: I1206 03:38:25.547939 4980 memory_manager.go:354] "RemoveStaleState removing state" podUID="965405e7-b022-4b55-945e-3b4e7adbc5e5" containerName="pruner" Dec 06 03:38:25 crc kubenswrapper[4980]: I1206 03:38:25.548004 4980 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Dec 06 03:38:25 crc kubenswrapper[4980]: I1206 03:38:25.548086 4980 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 06 03:38:25 crc kubenswrapper[4980]: I1206 03:38:25.548161 4980 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Dec 06 03:38:25 crc kubenswrapper[4980]: I1206 03:38:25.549956 4980 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Dec 06 03:38:25 crc kubenswrapper[4980]: I1206 03:38:25.550671 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 06 03:38:25 crc kubenswrapper[4980]: I1206 03:38:25.554828 4980 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="f4b27818a5e8e43d0dc095d08835c792" podUID="71bb4a3aecc4ba5b26c4b7318770ce13" Dec 06 03:38:25 crc kubenswrapper[4980]: I1206 03:38:25.586957 4980 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Dec 06 03:38:25 crc kubenswrapper[4980]: I1206 03:38:25.592962 4980 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-bl7hx" Dec 06 03:38:25 crc kubenswrapper[4980]: I1206 03:38:25.593256 4980 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-bl7hx" Dec 06 03:38:25 crc kubenswrapper[4980]: I1206 03:38:25.684807 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 06 03:38:25 crc kubenswrapper[4980]: I1206 03:38:25.684875 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 06 03:38:25 crc kubenswrapper[4980]: I1206 03:38:25.684912 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 06 03:38:25 crc kubenswrapper[4980]: I1206 03:38:25.684974 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 06 03:38:25 crc kubenswrapper[4980]: I1206 03:38:25.685003 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 06 03:38:25 crc kubenswrapper[4980]: I1206 03:38:25.685025 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 06 03:38:25 crc kubenswrapper[4980]: I1206 03:38:25.685053 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 06 03:38:25 crc kubenswrapper[4980]: I1206 03:38:25.685117 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 06 03:38:25 crc kubenswrapper[4980]: I1206 03:38:25.762872 4980 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Dec 06 03:38:25 crc kubenswrapper[4980]: I1206 03:38:25.766025 4980 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Dec 06 03:38:25 crc kubenswrapper[4980]: I1206 03:38:25.766716 4980 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="bafaa9670b05671fb57b658c2960bf3ee32c84e52f86a3a3bbab01d5a38ff0c4" exitCode=0 Dec 06 03:38:25 crc kubenswrapper[4980]: I1206 03:38:25.766738 4980 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="d40e0e7446f8416bf9da94fc53c01be937f9a51a11eb101157d2ddd4597720d2" exitCode=0 Dec 06 03:38:25 crc kubenswrapper[4980]: I1206 03:38:25.766747 4980 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="76cff8a2d652df38b04eefc053832baf4cbb5db778ff5e525b18c2825141b118" exitCode=0 Dec 06 03:38:25 crc kubenswrapper[4980]: I1206 03:38:25.766754 4980 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="206f6f19fe3197f5d6798cc86c5473bf2decb02a2860ee7ffb5efb5f60952e8c" exitCode=2 Dec 06 03:38:25 crc kubenswrapper[4980]: I1206 03:38:25.766848 4980 scope.go:117] "RemoveContainer" containerID="773a0bc2aab73ffc2717e987543defc9b786cc22300aa54b4ba4f3454504e331" Dec 06 03:38:25 crc kubenswrapper[4980]: I1206 03:38:25.772670 4980 generic.go:334] "Generic (PLEG): container finished" podID="652a7feb-30e3-4e9f-b1a9-c8ef201c658c" containerID="ffbb4c8c6980bde657d7596d37a800b08dd5ac764045bebabcafe2c50bd30ac6" exitCode=0 Dec 06 03:38:25 crc kubenswrapper[4980]: I1206 03:38:25.773572 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2vfzc" event={"ID":"652a7feb-30e3-4e9f-b1a9-c8ef201c658c","Type":"ContainerDied","Data":"ffbb4c8c6980bde657d7596d37a800b08dd5ac764045bebabcafe2c50bd30ac6"} Dec 06 03:38:25 crc kubenswrapper[4980]: I1206 03:38:25.775900 4980 status_manager.go:851] "Failed to get status for pod" podUID="652a7feb-30e3-4e9f-b1a9-c8ef201c658c" pod="openshift-marketplace/community-operators-2vfzc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-2vfzc\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 06 03:38:25 crc kubenswrapper[4980]: I1206 03:38:25.776223 4980 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 06 03:38:25 crc kubenswrapper[4980]: E1206 03:38:25.783409 4980 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/events\": dial tcp 38.102.83.184:6443: connect: connection refused" event="&Event{ObjectMeta:{community-operators-2vfzc.187e8326997e5a92 openshift-marketplace 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-marketplace,Name:community-operators-2vfzc,UID:652a7feb-30e3-4e9f-b1a9-c8ef201c658c,APIVersion:v1,ResourceVersion:28261,FieldPath:spec.containers{registry-server},},Reason:Pulling,Message:Pulling image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\",Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-06 03:38:25.783003794 +0000 UTC m=+265.027755065,LastTimestamp:2025-12-06 03:38:25.783003794 +0000 UTC m=+265.027755065,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Dec 06 03:38:25 crc kubenswrapper[4980]: I1206 03:38:25.788232 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 06 03:38:25 crc kubenswrapper[4980]: I1206 03:38:25.788368 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 06 03:38:25 crc kubenswrapper[4980]: I1206 03:38:25.788490 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 06 03:38:25 crc kubenswrapper[4980]: I1206 03:38:25.788295 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 06 03:38:25 crc kubenswrapper[4980]: I1206 03:38:25.788818 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 06 03:38:25 crc kubenswrapper[4980]: I1206 03:38:25.788651 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 06 03:38:25 crc kubenswrapper[4980]: I1206 03:38:25.788418 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 06 03:38:25 crc kubenswrapper[4980]: I1206 03:38:25.788718 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 06 03:38:25 crc kubenswrapper[4980]: I1206 03:38:25.789117 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 06 03:38:25 crc kubenswrapper[4980]: I1206 03:38:25.789261 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 06 03:38:25 crc kubenswrapper[4980]: I1206 03:38:25.789396 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 06 03:38:25 crc kubenswrapper[4980]: I1206 03:38:25.789201 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 06 03:38:25 crc kubenswrapper[4980]: I1206 03:38:25.789315 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 06 03:38:25 crc kubenswrapper[4980]: I1206 03:38:25.789444 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 06 03:38:25 crc kubenswrapper[4980]: I1206 03:38:25.790020 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 06 03:38:25 crc kubenswrapper[4980]: I1206 03:38:25.790176 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 06 03:38:25 crc kubenswrapper[4980]: I1206 03:38:25.887011 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 06 03:38:26 crc kubenswrapper[4980]: W1206 03:38:26.012878 4980 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf85e55b1a89d02b0cb034b1ea31ed45a.slice/crio-6ecfc7e1717f7cb247e7ea4547233bc4cc17bb1654e4b007d68b6b941d7e3911 WatchSource:0}: Error finding container 6ecfc7e1717f7cb247e7ea4547233bc4cc17bb1654e4b007d68b6b941d7e3911: Status 404 returned error can't find the container with id 6ecfc7e1717f7cb247e7ea4547233bc4cc17bb1654e4b007d68b6b941d7e3911 Dec 06 03:38:26 crc kubenswrapper[4980]: I1206 03:38:26.670657 4980 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-bl7hx" podUID="56a102c7-f341-4076-89dc-44428e77c164" containerName="registry-server" probeResult="failure" output=< Dec 06 03:38:26 crc kubenswrapper[4980]: timeout: failed to connect service ":50051" within 1s Dec 06 03:38:26 crc kubenswrapper[4980]: > Dec 06 03:38:26 crc kubenswrapper[4980]: I1206 03:38:26.807754 4980 generic.go:334] "Generic (PLEG): container finished" podID="43112506-9fdd-4b19-a5a5-caf864c09774" containerID="beb2e6aabb313ede1d7348363d791410e4e40f483c9f4bfceec700e83315d0b9" exitCode=0 Dec 06 03:38:26 crc kubenswrapper[4980]: I1206 03:38:26.807837 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-m9vl6" event={"ID":"43112506-9fdd-4b19-a5a5-caf864c09774","Type":"ContainerDied","Data":"beb2e6aabb313ede1d7348363d791410e4e40f483c9f4bfceec700e83315d0b9"} Dec 06 03:38:26 crc kubenswrapper[4980]: I1206 03:38:26.808621 4980 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 06 03:38:26 crc kubenswrapper[4980]: I1206 03:38:26.809009 4980 status_manager.go:851] "Failed to get status for pod" podUID="652a7feb-30e3-4e9f-b1a9-c8ef201c658c" pod="openshift-marketplace/community-operators-2vfzc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-2vfzc\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 06 03:38:26 crc kubenswrapper[4980]: I1206 03:38:26.809524 4980 status_manager.go:851] "Failed to get status for pod" podUID="43112506-9fdd-4b19-a5a5-caf864c09774" pod="openshift-marketplace/redhat-operators-m9vl6" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-m9vl6\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 06 03:38:26 crc kubenswrapper[4980]: I1206 03:38:26.810146 4980 generic.go:334] "Generic (PLEG): container finished" podID="dee63a36-cd1b-410c-b329-bd7966b1d38d" containerID="61478a7d59badbae7bb924abe9d3ca7dd0623837917b65a27b2d83d7521a73cb" exitCode=0 Dec 06 03:38:26 crc kubenswrapper[4980]: I1206 03:38:26.810219 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"dee63a36-cd1b-410c-b329-bd7966b1d38d","Type":"ContainerDied","Data":"61478a7d59badbae7bb924abe9d3ca7dd0623837917b65a27b2d83d7521a73cb"} Dec 06 03:38:26 crc kubenswrapper[4980]: I1206 03:38:26.810657 4980 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 06 03:38:26 crc kubenswrapper[4980]: I1206 03:38:26.810899 4980 status_manager.go:851] "Failed to get status for pod" podUID="652a7feb-30e3-4e9f-b1a9-c8ef201c658c" pod="openshift-marketplace/community-operators-2vfzc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-2vfzc\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 06 03:38:26 crc kubenswrapper[4980]: I1206 03:38:26.811241 4980 status_manager.go:851] "Failed to get status for pod" podUID="43112506-9fdd-4b19-a5a5-caf864c09774" pod="openshift-marketplace/redhat-operators-m9vl6" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-m9vl6\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 06 03:38:26 crc kubenswrapper[4980]: I1206 03:38:26.811898 4980 status_manager.go:851] "Failed to get status for pod" podUID="dee63a36-cd1b-410c-b329-bd7966b1d38d" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 06 03:38:26 crc kubenswrapper[4980]: I1206 03:38:26.812226 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" event={"ID":"f85e55b1a89d02b0cb034b1ea31ed45a","Type":"ContainerStarted","Data":"86f06d4f20478f6fc3524bcb449288b21a827bc591d313aecbb0772ed610d129"} Dec 06 03:38:26 crc kubenswrapper[4980]: I1206 03:38:26.812262 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" event={"ID":"f85e55b1a89d02b0cb034b1ea31ed45a","Type":"ContainerStarted","Data":"6ecfc7e1717f7cb247e7ea4547233bc4cc17bb1654e4b007d68b6b941d7e3911"} Dec 06 03:38:26 crc kubenswrapper[4980]: I1206 03:38:26.813178 4980 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 06 03:38:26 crc kubenswrapper[4980]: I1206 03:38:26.813492 4980 status_manager.go:851] "Failed to get status for pod" podUID="652a7feb-30e3-4e9f-b1a9-c8ef201c658c" pod="openshift-marketplace/community-operators-2vfzc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-2vfzc\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 06 03:38:26 crc kubenswrapper[4980]: I1206 03:38:26.813775 4980 status_manager.go:851] "Failed to get status for pod" podUID="43112506-9fdd-4b19-a5a5-caf864c09774" pod="openshift-marketplace/redhat-operators-m9vl6" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-m9vl6\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 06 03:38:26 crc kubenswrapper[4980]: I1206 03:38:26.814033 4980 status_manager.go:851] "Failed to get status for pod" podUID="dee63a36-cd1b-410c-b329-bd7966b1d38d" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 06 03:38:26 crc kubenswrapper[4980]: I1206 03:38:26.817784 4980 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Dec 06 03:38:27 crc kubenswrapper[4980]: I1206 03:38:27.827316 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-m9vl6" event={"ID":"43112506-9fdd-4b19-a5a5-caf864c09774","Type":"ContainerStarted","Data":"814e5ffce677f097050c8e056ca0ebc60c478a53dbfdeffbc4ad76fabbd8622b"} Dec 06 03:38:27 crc kubenswrapper[4980]: I1206 03:38:27.829504 4980 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 06 03:38:27 crc kubenswrapper[4980]: I1206 03:38:27.829632 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2vfzc" event={"ID":"652a7feb-30e3-4e9f-b1a9-c8ef201c658c","Type":"ContainerStarted","Data":"f0efc2ce0c1863787e77589a8a0afc2f9e4fdb3ce974b5bcac67705360f28d24"} Dec 06 03:38:27 crc kubenswrapper[4980]: I1206 03:38:27.829783 4980 status_manager.go:851] "Failed to get status for pod" podUID="652a7feb-30e3-4e9f-b1a9-c8ef201c658c" pod="openshift-marketplace/community-operators-2vfzc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-2vfzc\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 06 03:38:27 crc kubenswrapper[4980]: I1206 03:38:27.830338 4980 status_manager.go:851] "Failed to get status for pod" podUID="43112506-9fdd-4b19-a5a5-caf864c09774" pod="openshift-marketplace/redhat-operators-m9vl6" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-m9vl6\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 06 03:38:27 crc kubenswrapper[4980]: I1206 03:38:27.830621 4980 status_manager.go:851] "Failed to get status for pod" podUID="dee63a36-cd1b-410c-b329-bd7966b1d38d" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 06 03:38:27 crc kubenswrapper[4980]: I1206 03:38:27.830927 4980 status_manager.go:851] "Failed to get status for pod" podUID="dee63a36-cd1b-410c-b329-bd7966b1d38d" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 06 03:38:27 crc kubenswrapper[4980]: I1206 03:38:27.831178 4980 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 06 03:38:27 crc kubenswrapper[4980]: I1206 03:38:27.831734 4980 status_manager.go:851] "Failed to get status for pod" podUID="652a7feb-30e3-4e9f-b1a9-c8ef201c658c" pod="openshift-marketplace/community-operators-2vfzc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-2vfzc\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 06 03:38:27 crc kubenswrapper[4980]: I1206 03:38:27.831981 4980 status_manager.go:851] "Failed to get status for pod" podUID="43112506-9fdd-4b19-a5a5-caf864c09774" pod="openshift-marketplace/redhat-operators-m9vl6" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-m9vl6\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 06 03:38:28 crc kubenswrapper[4980]: I1206 03:38:28.048912 4980 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Dec 06 03:38:28 crc kubenswrapper[4980]: I1206 03:38:28.050120 4980 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 06 03:38:28 crc kubenswrapper[4980]: I1206 03:38:28.054705 4980 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 06 03:38:28 crc kubenswrapper[4980]: I1206 03:38:28.055230 4980 status_manager.go:851] "Failed to get status for pod" podUID="43112506-9fdd-4b19-a5a5-caf864c09774" pod="openshift-marketplace/redhat-operators-m9vl6" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-m9vl6\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 06 03:38:28 crc kubenswrapper[4980]: I1206 03:38:28.055539 4980 status_manager.go:851] "Failed to get status for pod" podUID="652a7feb-30e3-4e9f-b1a9-c8ef201c658c" pod="openshift-marketplace/community-operators-2vfzc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-2vfzc\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 06 03:38:28 crc kubenswrapper[4980]: I1206 03:38:28.055858 4980 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 06 03:38:28 crc kubenswrapper[4980]: I1206 03:38:28.056128 4980 status_manager.go:851] "Failed to get status for pod" podUID="dee63a36-cd1b-410c-b329-bd7966b1d38d" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 06 03:38:28 crc kubenswrapper[4980]: I1206 03:38:28.135725 4980 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Dec 06 03:38:28 crc kubenswrapper[4980]: I1206 03:38:28.136181 4980 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 06 03:38:28 crc kubenswrapper[4980]: I1206 03:38:28.136412 4980 status_manager.go:851] "Failed to get status for pod" podUID="652a7feb-30e3-4e9f-b1a9-c8ef201c658c" pod="openshift-marketplace/community-operators-2vfzc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-2vfzc\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 06 03:38:28 crc kubenswrapper[4980]: I1206 03:38:28.136647 4980 status_manager.go:851] "Failed to get status for pod" podUID="43112506-9fdd-4b19-a5a5-caf864c09774" pod="openshift-marketplace/redhat-operators-m9vl6" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-m9vl6\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 06 03:38:28 crc kubenswrapper[4980]: I1206 03:38:28.136805 4980 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 06 03:38:28 crc kubenswrapper[4980]: I1206 03:38:28.136985 4980 status_manager.go:851] "Failed to get status for pod" podUID="dee63a36-cd1b-410c-b329-bd7966b1d38d" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 06 03:38:28 crc kubenswrapper[4980]: I1206 03:38:28.187202 4980 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-authentication/oauth-openshift-558db77b4-pmmcw" podUID="a42aad68-42cf-4264-9c94-8678b0eea1b8" containerName="oauth-openshift" containerID="cri-o://679b95f34ba069aebe784949cf4ef7693988ae7acaa83b3bd44b881022728b30" gracePeriod=15 Dec 06 03:38:28 crc kubenswrapper[4980]: I1206 03:38:28.237803 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Dec 06 03:38:28 crc kubenswrapper[4980]: I1206 03:38:28.238235 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/dee63a36-cd1b-410c-b329-bd7966b1d38d-kube-api-access\") pod \"dee63a36-cd1b-410c-b329-bd7966b1d38d\" (UID: \"dee63a36-cd1b-410c-b329-bd7966b1d38d\") " Dec 06 03:38:28 crc kubenswrapper[4980]: I1206 03:38:28.238278 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/dee63a36-cd1b-410c-b329-bd7966b1d38d-var-lock\") pod \"dee63a36-cd1b-410c-b329-bd7966b1d38d\" (UID: \"dee63a36-cd1b-410c-b329-bd7966b1d38d\") " Dec 06 03:38:28 crc kubenswrapper[4980]: I1206 03:38:28.238315 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Dec 06 03:38:28 crc kubenswrapper[4980]: I1206 03:38:28.238361 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/dee63a36-cd1b-410c-b329-bd7966b1d38d-kubelet-dir\") pod \"dee63a36-cd1b-410c-b329-bd7966b1d38d\" (UID: \"dee63a36-cd1b-410c-b329-bd7966b1d38d\") " Dec 06 03:38:28 crc kubenswrapper[4980]: I1206 03:38:28.238384 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Dec 06 03:38:28 crc kubenswrapper[4980]: I1206 03:38:28.238006 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 03:38:28 crc kubenswrapper[4980]: I1206 03:38:28.238735 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/dee63a36-cd1b-410c-b329-bd7966b1d38d-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "dee63a36-cd1b-410c-b329-bd7966b1d38d" (UID: "dee63a36-cd1b-410c-b329-bd7966b1d38d"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 03:38:28 crc kubenswrapper[4980]: I1206 03:38:28.238698 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir" (OuterVolumeSpecName: "resource-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 03:38:28 crc kubenswrapper[4980]: I1206 03:38:28.238717 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir" (OuterVolumeSpecName: "cert-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "cert-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 03:38:28 crc kubenswrapper[4980]: I1206 03:38:28.238718 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/dee63a36-cd1b-410c-b329-bd7966b1d38d-var-lock" (OuterVolumeSpecName: "var-lock") pod "dee63a36-cd1b-410c-b329-bd7966b1d38d" (UID: "dee63a36-cd1b-410c-b329-bd7966b1d38d"). InnerVolumeSpecName "var-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 03:38:28 crc kubenswrapper[4980]: I1206 03:38:28.245355 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dee63a36-cd1b-410c-b329-bd7966b1d38d-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "dee63a36-cd1b-410c-b329-bd7966b1d38d" (UID: "dee63a36-cd1b-410c-b329-bd7966b1d38d"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 03:38:28 crc kubenswrapper[4980]: I1206 03:38:28.339482 4980 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/dee63a36-cd1b-410c-b329-bd7966b1d38d-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 06 03:38:28 crc kubenswrapper[4980]: I1206 03:38:28.339533 4980 reconciler_common.go:293] "Volume detached for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/dee63a36-cd1b-410c-b329-bd7966b1d38d-var-lock\") on node \"crc\" DevicePath \"\"" Dec 06 03:38:28 crc kubenswrapper[4980]: I1206 03:38:28.339545 4980 reconciler_common.go:293] "Volume detached for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") on node \"crc\" DevicePath \"\"" Dec 06 03:38:28 crc kubenswrapper[4980]: I1206 03:38:28.339556 4980 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/dee63a36-cd1b-410c-b329-bd7966b1d38d-kubelet-dir\") on node \"crc\" DevicePath \"\"" Dec 06 03:38:28 crc kubenswrapper[4980]: I1206 03:38:28.339565 4980 reconciler_common.go:293] "Volume detached for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") on node \"crc\" DevicePath \"\"" Dec 06 03:38:28 crc kubenswrapper[4980]: I1206 03:38:28.339576 4980 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") on node \"crc\" DevicePath \"\"" Dec 06 03:38:28 crc kubenswrapper[4980]: I1206 03:38:28.560265 4980 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-pmmcw" Dec 06 03:38:28 crc kubenswrapper[4980]: I1206 03:38:28.560952 4980 status_manager.go:851] "Failed to get status for pod" podUID="a42aad68-42cf-4264-9c94-8678b0eea1b8" pod="openshift-authentication/oauth-openshift-558db77b4-pmmcw" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/pods/oauth-openshift-558db77b4-pmmcw\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 06 03:38:28 crc kubenswrapper[4980]: I1206 03:38:28.561232 4980 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 06 03:38:28 crc kubenswrapper[4980]: I1206 03:38:28.561424 4980 status_manager.go:851] "Failed to get status for pod" podUID="652a7feb-30e3-4e9f-b1a9-c8ef201c658c" pod="openshift-marketplace/community-operators-2vfzc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-2vfzc\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 06 03:38:28 crc kubenswrapper[4980]: I1206 03:38:28.561635 4980 status_manager.go:851] "Failed to get status for pod" podUID="43112506-9fdd-4b19-a5a5-caf864c09774" pod="openshift-marketplace/redhat-operators-m9vl6" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-m9vl6\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 06 03:38:28 crc kubenswrapper[4980]: I1206 03:38:28.561834 4980 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 06 03:38:28 crc kubenswrapper[4980]: I1206 03:38:28.562022 4980 status_manager.go:851] "Failed to get status for pod" podUID="dee63a36-cd1b-410c-b329-bd7966b1d38d" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 06 03:38:28 crc kubenswrapper[4980]: I1206 03:38:28.744018 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/a42aad68-42cf-4264-9c94-8678b0eea1b8-v4-0-config-system-ocp-branding-template\") pod \"a42aad68-42cf-4264-9c94-8678b0eea1b8\" (UID: \"a42aad68-42cf-4264-9c94-8678b0eea1b8\") " Dec 06 03:38:28 crc kubenswrapper[4980]: I1206 03:38:28.744093 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/a42aad68-42cf-4264-9c94-8678b0eea1b8-v4-0-config-user-template-provider-selection\") pod \"a42aad68-42cf-4264-9c94-8678b0eea1b8\" (UID: \"a42aad68-42cf-4264-9c94-8678b0eea1b8\") " Dec 06 03:38:28 crc kubenswrapper[4980]: I1206 03:38:28.744129 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/a42aad68-42cf-4264-9c94-8678b0eea1b8-v4-0-config-system-router-certs\") pod \"a42aad68-42cf-4264-9c94-8678b0eea1b8\" (UID: \"a42aad68-42cf-4264-9c94-8678b0eea1b8\") " Dec 06 03:38:28 crc kubenswrapper[4980]: I1206 03:38:28.744164 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/a42aad68-42cf-4264-9c94-8678b0eea1b8-v4-0-config-user-idp-0-file-data\") pod \"a42aad68-42cf-4264-9c94-8678b0eea1b8\" (UID: \"a42aad68-42cf-4264-9c94-8678b0eea1b8\") " Dec 06 03:38:28 crc kubenswrapper[4980]: I1206 03:38:28.744193 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kzm64\" (UniqueName: \"kubernetes.io/projected/a42aad68-42cf-4264-9c94-8678b0eea1b8-kube-api-access-kzm64\") pod \"a42aad68-42cf-4264-9c94-8678b0eea1b8\" (UID: \"a42aad68-42cf-4264-9c94-8678b0eea1b8\") " Dec 06 03:38:28 crc kubenswrapper[4980]: I1206 03:38:28.744208 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/a42aad68-42cf-4264-9c94-8678b0eea1b8-v4-0-config-user-template-error\") pod \"a42aad68-42cf-4264-9c94-8678b0eea1b8\" (UID: \"a42aad68-42cf-4264-9c94-8678b0eea1b8\") " Dec 06 03:38:28 crc kubenswrapper[4980]: I1206 03:38:28.744231 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a42aad68-42cf-4264-9c94-8678b0eea1b8-v4-0-config-system-trusted-ca-bundle\") pod \"a42aad68-42cf-4264-9c94-8678b0eea1b8\" (UID: \"a42aad68-42cf-4264-9c94-8678b0eea1b8\") " Dec 06 03:38:28 crc kubenswrapper[4980]: I1206 03:38:28.744266 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/a42aad68-42cf-4264-9c94-8678b0eea1b8-v4-0-config-system-session\") pod \"a42aad68-42cf-4264-9c94-8678b0eea1b8\" (UID: \"a42aad68-42cf-4264-9c94-8678b0eea1b8\") " Dec 06 03:38:28 crc kubenswrapper[4980]: I1206 03:38:28.744294 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/a42aad68-42cf-4264-9c94-8678b0eea1b8-v4-0-config-system-service-ca\") pod \"a42aad68-42cf-4264-9c94-8678b0eea1b8\" (UID: \"a42aad68-42cf-4264-9c94-8678b0eea1b8\") " Dec 06 03:38:28 crc kubenswrapper[4980]: I1206 03:38:28.744313 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/a42aad68-42cf-4264-9c94-8678b0eea1b8-audit-policies\") pod \"a42aad68-42cf-4264-9c94-8678b0eea1b8\" (UID: \"a42aad68-42cf-4264-9c94-8678b0eea1b8\") " Dec 06 03:38:28 crc kubenswrapper[4980]: I1206 03:38:28.744334 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/a42aad68-42cf-4264-9c94-8678b0eea1b8-v4-0-config-system-serving-cert\") pod \"a42aad68-42cf-4264-9c94-8678b0eea1b8\" (UID: \"a42aad68-42cf-4264-9c94-8678b0eea1b8\") " Dec 06 03:38:28 crc kubenswrapper[4980]: I1206 03:38:28.744351 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/a42aad68-42cf-4264-9c94-8678b0eea1b8-v4-0-config-user-template-login\") pod \"a42aad68-42cf-4264-9c94-8678b0eea1b8\" (UID: \"a42aad68-42cf-4264-9c94-8678b0eea1b8\") " Dec 06 03:38:28 crc kubenswrapper[4980]: I1206 03:38:28.744389 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/a42aad68-42cf-4264-9c94-8678b0eea1b8-audit-dir\") pod \"a42aad68-42cf-4264-9c94-8678b0eea1b8\" (UID: \"a42aad68-42cf-4264-9c94-8678b0eea1b8\") " Dec 06 03:38:28 crc kubenswrapper[4980]: I1206 03:38:28.744406 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/a42aad68-42cf-4264-9c94-8678b0eea1b8-v4-0-config-system-cliconfig\") pod \"a42aad68-42cf-4264-9c94-8678b0eea1b8\" (UID: \"a42aad68-42cf-4264-9c94-8678b0eea1b8\") " Dec 06 03:38:28 crc kubenswrapper[4980]: I1206 03:38:28.745357 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a42aad68-42cf-4264-9c94-8678b0eea1b8-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "a42aad68-42cf-4264-9c94-8678b0eea1b8" (UID: "a42aad68-42cf-4264-9c94-8678b0eea1b8"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 03:38:28 crc kubenswrapper[4980]: I1206 03:38:28.745925 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a42aad68-42cf-4264-9c94-8678b0eea1b8-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "a42aad68-42cf-4264-9c94-8678b0eea1b8" (UID: "a42aad68-42cf-4264-9c94-8678b0eea1b8"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 03:38:28 crc kubenswrapper[4980]: I1206 03:38:28.747134 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a42aad68-42cf-4264-9c94-8678b0eea1b8-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "a42aad68-42cf-4264-9c94-8678b0eea1b8" (UID: "a42aad68-42cf-4264-9c94-8678b0eea1b8"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 03:38:28 crc kubenswrapper[4980]: I1206 03:38:28.747198 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a42aad68-42cf-4264-9c94-8678b0eea1b8-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "a42aad68-42cf-4264-9c94-8678b0eea1b8" (UID: "a42aad68-42cf-4264-9c94-8678b0eea1b8"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 03:38:28 crc kubenswrapper[4980]: I1206 03:38:28.747476 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a42aad68-42cf-4264-9c94-8678b0eea1b8-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "a42aad68-42cf-4264-9c94-8678b0eea1b8" (UID: "a42aad68-42cf-4264-9c94-8678b0eea1b8"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 03:38:28 crc kubenswrapper[4980]: I1206 03:38:28.748183 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a42aad68-42cf-4264-9c94-8678b0eea1b8-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "a42aad68-42cf-4264-9c94-8678b0eea1b8" (UID: "a42aad68-42cf-4264-9c94-8678b0eea1b8"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 03:38:28 crc kubenswrapper[4980]: I1206 03:38:28.748459 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a42aad68-42cf-4264-9c94-8678b0eea1b8-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "a42aad68-42cf-4264-9c94-8678b0eea1b8" (UID: "a42aad68-42cf-4264-9c94-8678b0eea1b8"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 03:38:28 crc kubenswrapper[4980]: I1206 03:38:28.749216 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a42aad68-42cf-4264-9c94-8678b0eea1b8-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "a42aad68-42cf-4264-9c94-8678b0eea1b8" (UID: "a42aad68-42cf-4264-9c94-8678b0eea1b8"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 03:38:28 crc kubenswrapper[4980]: I1206 03:38:28.749349 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a42aad68-42cf-4264-9c94-8678b0eea1b8-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "a42aad68-42cf-4264-9c94-8678b0eea1b8" (UID: "a42aad68-42cf-4264-9c94-8678b0eea1b8"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 03:38:28 crc kubenswrapper[4980]: I1206 03:38:28.751329 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a42aad68-42cf-4264-9c94-8678b0eea1b8-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "a42aad68-42cf-4264-9c94-8678b0eea1b8" (UID: "a42aad68-42cf-4264-9c94-8678b0eea1b8"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 03:38:28 crc kubenswrapper[4980]: I1206 03:38:28.751343 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a42aad68-42cf-4264-9c94-8678b0eea1b8-kube-api-access-kzm64" (OuterVolumeSpecName: "kube-api-access-kzm64") pod "a42aad68-42cf-4264-9c94-8678b0eea1b8" (UID: "a42aad68-42cf-4264-9c94-8678b0eea1b8"). InnerVolumeSpecName "kube-api-access-kzm64". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 03:38:28 crc kubenswrapper[4980]: I1206 03:38:28.751560 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a42aad68-42cf-4264-9c94-8678b0eea1b8-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "a42aad68-42cf-4264-9c94-8678b0eea1b8" (UID: "a42aad68-42cf-4264-9c94-8678b0eea1b8"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 03:38:28 crc kubenswrapper[4980]: I1206 03:38:28.751709 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a42aad68-42cf-4264-9c94-8678b0eea1b8-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "a42aad68-42cf-4264-9c94-8678b0eea1b8" (UID: "a42aad68-42cf-4264-9c94-8678b0eea1b8"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 03:38:28 crc kubenswrapper[4980]: I1206 03:38:28.752608 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a42aad68-42cf-4264-9c94-8678b0eea1b8-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "a42aad68-42cf-4264-9c94-8678b0eea1b8" (UID: "a42aad68-42cf-4264-9c94-8678b0eea1b8"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 03:38:28 crc kubenswrapper[4980]: I1206 03:38:28.836414 4980 generic.go:334] "Generic (PLEG): container finished" podID="a42aad68-42cf-4264-9c94-8678b0eea1b8" containerID="679b95f34ba069aebe784949cf4ef7693988ae7acaa83b3bd44b881022728b30" exitCode=0 Dec 06 03:38:28 crc kubenswrapper[4980]: I1206 03:38:28.836575 4980 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-pmmcw" Dec 06 03:38:28 crc kubenswrapper[4980]: I1206 03:38:28.836856 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-pmmcw" event={"ID":"a42aad68-42cf-4264-9c94-8678b0eea1b8","Type":"ContainerDied","Data":"679b95f34ba069aebe784949cf4ef7693988ae7acaa83b3bd44b881022728b30"} Dec 06 03:38:28 crc kubenswrapper[4980]: I1206 03:38:28.836911 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-pmmcw" event={"ID":"a42aad68-42cf-4264-9c94-8678b0eea1b8","Type":"ContainerDied","Data":"95185ed7e417885129fd12417bd2f6923e4f65a41567a2a74c075efe7067b280"} Dec 06 03:38:28 crc kubenswrapper[4980]: I1206 03:38:28.836942 4980 scope.go:117] "RemoveContainer" containerID="679b95f34ba069aebe784949cf4ef7693988ae7acaa83b3bd44b881022728b30" Dec 06 03:38:28 crc kubenswrapper[4980]: I1206 03:38:28.838164 4980 status_manager.go:851] "Failed to get status for pod" podUID="a42aad68-42cf-4264-9c94-8678b0eea1b8" pod="openshift-authentication/oauth-openshift-558db77b4-pmmcw" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/pods/oauth-openshift-558db77b4-pmmcw\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 06 03:38:28 crc kubenswrapper[4980]: I1206 03:38:28.838375 4980 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 06 03:38:28 crc kubenswrapper[4980]: I1206 03:38:28.838610 4980 status_manager.go:851] "Failed to get status for pod" podUID="43112506-9fdd-4b19-a5a5-caf864c09774" pod="openshift-marketplace/redhat-operators-m9vl6" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-m9vl6\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 06 03:38:28 crc kubenswrapper[4980]: I1206 03:38:28.838845 4980 status_manager.go:851] "Failed to get status for pod" podUID="652a7feb-30e3-4e9f-b1a9-c8ef201c658c" pod="openshift-marketplace/community-operators-2vfzc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-2vfzc\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 06 03:38:28 crc kubenswrapper[4980]: I1206 03:38:28.839050 4980 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 06 03:38:28 crc kubenswrapper[4980]: I1206 03:38:28.839248 4980 status_manager.go:851] "Failed to get status for pod" podUID="dee63a36-cd1b-410c-b329-bd7966b1d38d" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 06 03:38:28 crc kubenswrapper[4980]: I1206 03:38:28.839574 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"dee63a36-cd1b-410c-b329-bd7966b1d38d","Type":"ContainerDied","Data":"1bb044f31576bca8e79d7140ce8c441266db23773b959fe99d8cc5a6e40f2462"} Dec 06 03:38:28 crc kubenswrapper[4980]: I1206 03:38:28.839623 4980 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1bb044f31576bca8e79d7140ce8c441266db23773b959fe99d8cc5a6e40f2462" Dec 06 03:38:28 crc kubenswrapper[4980]: I1206 03:38:28.839700 4980 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Dec 06 03:38:28 crc kubenswrapper[4980]: I1206 03:38:28.845889 4980 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/a42aad68-42cf-4264-9c94-8678b0eea1b8-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Dec 06 03:38:28 crc kubenswrapper[4980]: I1206 03:38:28.845930 4980 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/a42aad68-42cf-4264-9c94-8678b0eea1b8-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Dec 06 03:38:28 crc kubenswrapper[4980]: I1206 03:38:28.845944 4980 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/a42aad68-42cf-4264-9c94-8678b0eea1b8-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Dec 06 03:38:28 crc kubenswrapper[4980]: I1206 03:38:28.845960 4980 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/a42aad68-42cf-4264-9c94-8678b0eea1b8-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Dec 06 03:38:28 crc kubenswrapper[4980]: I1206 03:38:28.845972 4980 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/a42aad68-42cf-4264-9c94-8678b0eea1b8-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Dec 06 03:38:28 crc kubenswrapper[4980]: I1206 03:38:28.845984 4980 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/a42aad68-42cf-4264-9c94-8678b0eea1b8-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Dec 06 03:38:28 crc kubenswrapper[4980]: I1206 03:38:28.845998 4980 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kzm64\" (UniqueName: \"kubernetes.io/projected/a42aad68-42cf-4264-9c94-8678b0eea1b8-kube-api-access-kzm64\") on node \"crc\" DevicePath \"\"" Dec 06 03:38:28 crc kubenswrapper[4980]: I1206 03:38:28.846010 4980 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a42aad68-42cf-4264-9c94-8678b0eea1b8-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 03:38:28 crc kubenswrapper[4980]: I1206 03:38:28.846024 4980 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/a42aad68-42cf-4264-9c94-8678b0eea1b8-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Dec 06 03:38:28 crc kubenswrapper[4980]: I1206 03:38:28.846035 4980 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/a42aad68-42cf-4264-9c94-8678b0eea1b8-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Dec 06 03:38:28 crc kubenswrapper[4980]: I1206 03:38:28.846047 4980 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/a42aad68-42cf-4264-9c94-8678b0eea1b8-audit-policies\") on node \"crc\" DevicePath \"\"" Dec 06 03:38:28 crc kubenswrapper[4980]: I1206 03:38:28.846059 4980 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/a42aad68-42cf-4264-9c94-8678b0eea1b8-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 06 03:38:28 crc kubenswrapper[4980]: I1206 03:38:28.846073 4980 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/a42aad68-42cf-4264-9c94-8678b0eea1b8-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Dec 06 03:38:28 crc kubenswrapper[4980]: I1206 03:38:28.846085 4980 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/a42aad68-42cf-4264-9c94-8678b0eea1b8-audit-dir\") on node \"crc\" DevicePath \"\"" Dec 06 03:38:28 crc kubenswrapper[4980]: I1206 03:38:28.846704 4980 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Dec 06 03:38:28 crc kubenswrapper[4980]: I1206 03:38:28.847897 4980 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="c28b6895b14e73f4f3eb7a8ce93076c05092ba9b58ea0e3e879c65bcfa227a21" exitCode=0 Dec 06 03:38:28 crc kubenswrapper[4980]: I1206 03:38:28.848025 4980 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 06 03:38:28 crc kubenswrapper[4980]: I1206 03:38:28.857794 4980 status_manager.go:851] "Failed to get status for pod" podUID="a42aad68-42cf-4264-9c94-8678b0eea1b8" pod="openshift-authentication/oauth-openshift-558db77b4-pmmcw" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/pods/oauth-openshift-558db77b4-pmmcw\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 06 03:38:28 crc kubenswrapper[4980]: I1206 03:38:28.858117 4980 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 06 03:38:28 crc kubenswrapper[4980]: I1206 03:38:28.858435 4980 status_manager.go:851] "Failed to get status for pod" podUID="652a7feb-30e3-4e9f-b1a9-c8ef201c658c" pod="openshift-marketplace/community-operators-2vfzc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-2vfzc\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 06 03:38:28 crc kubenswrapper[4980]: I1206 03:38:28.858606 4980 status_manager.go:851] "Failed to get status for pod" podUID="43112506-9fdd-4b19-a5a5-caf864c09774" pod="openshift-marketplace/redhat-operators-m9vl6" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-m9vl6\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 06 03:38:28 crc kubenswrapper[4980]: I1206 03:38:28.858745 4980 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 06 03:38:28 crc kubenswrapper[4980]: I1206 03:38:28.858912 4980 status_manager.go:851] "Failed to get status for pod" podUID="dee63a36-cd1b-410c-b329-bd7966b1d38d" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 06 03:38:28 crc kubenswrapper[4980]: I1206 03:38:28.859150 4980 status_manager.go:851] "Failed to get status for pod" podUID="dee63a36-cd1b-410c-b329-bd7966b1d38d" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 06 03:38:28 crc kubenswrapper[4980]: I1206 03:38:28.859237 4980 scope.go:117] "RemoveContainer" containerID="679b95f34ba069aebe784949cf4ef7693988ae7acaa83b3bd44b881022728b30" Dec 06 03:38:28 crc kubenswrapper[4980]: I1206 03:38:28.859323 4980 status_manager.go:851] "Failed to get status for pod" podUID="a42aad68-42cf-4264-9c94-8678b0eea1b8" pod="openshift-authentication/oauth-openshift-558db77b4-pmmcw" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/pods/oauth-openshift-558db77b4-pmmcw\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 06 03:38:28 crc kubenswrapper[4980]: I1206 03:38:28.859585 4980 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 06 03:38:28 crc kubenswrapper[4980]: I1206 03:38:28.859798 4980 status_manager.go:851] "Failed to get status for pod" podUID="652a7feb-30e3-4e9f-b1a9-c8ef201c658c" pod="openshift-marketplace/community-operators-2vfzc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-2vfzc\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 06 03:38:28 crc kubenswrapper[4980]: E1206 03:38:28.859865 4980 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"679b95f34ba069aebe784949cf4ef7693988ae7acaa83b3bd44b881022728b30\": container with ID starting with 679b95f34ba069aebe784949cf4ef7693988ae7acaa83b3bd44b881022728b30 not found: ID does not exist" containerID="679b95f34ba069aebe784949cf4ef7693988ae7acaa83b3bd44b881022728b30" Dec 06 03:38:28 crc kubenswrapper[4980]: I1206 03:38:28.859894 4980 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"679b95f34ba069aebe784949cf4ef7693988ae7acaa83b3bd44b881022728b30"} err="failed to get container status \"679b95f34ba069aebe784949cf4ef7693988ae7acaa83b3bd44b881022728b30\": rpc error: code = NotFound desc = could not find container \"679b95f34ba069aebe784949cf4ef7693988ae7acaa83b3bd44b881022728b30\": container with ID starting with 679b95f34ba069aebe784949cf4ef7693988ae7acaa83b3bd44b881022728b30 not found: ID does not exist" Dec 06 03:38:28 crc kubenswrapper[4980]: I1206 03:38:28.859918 4980 scope.go:117] "RemoveContainer" containerID="bafaa9670b05671fb57b658c2960bf3ee32c84e52f86a3a3bbab01d5a38ff0c4" Dec 06 03:38:28 crc kubenswrapper[4980]: I1206 03:38:28.861322 4980 status_manager.go:851] "Failed to get status for pod" podUID="43112506-9fdd-4b19-a5a5-caf864c09774" pod="openshift-marketplace/redhat-operators-m9vl6" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-m9vl6\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 06 03:38:28 crc kubenswrapper[4980]: I1206 03:38:28.861638 4980 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 06 03:38:28 crc kubenswrapper[4980]: I1206 03:38:28.876651 4980 status_manager.go:851] "Failed to get status for pod" podUID="652a7feb-30e3-4e9f-b1a9-c8ef201c658c" pod="openshift-marketplace/community-operators-2vfzc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-2vfzc\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 06 03:38:28 crc kubenswrapper[4980]: I1206 03:38:28.877218 4980 status_manager.go:851] "Failed to get status for pod" podUID="43112506-9fdd-4b19-a5a5-caf864c09774" pod="openshift-marketplace/redhat-operators-m9vl6" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-m9vl6\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 06 03:38:28 crc kubenswrapper[4980]: I1206 03:38:28.877405 4980 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 06 03:38:28 crc kubenswrapper[4980]: I1206 03:38:28.877574 4980 status_manager.go:851] "Failed to get status for pod" podUID="dee63a36-cd1b-410c-b329-bd7966b1d38d" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 06 03:38:28 crc kubenswrapper[4980]: I1206 03:38:28.877735 4980 status_manager.go:851] "Failed to get status for pod" podUID="a42aad68-42cf-4264-9c94-8678b0eea1b8" pod="openshift-authentication/oauth-openshift-558db77b4-pmmcw" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/pods/oauth-openshift-558db77b4-pmmcw\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 06 03:38:28 crc kubenswrapper[4980]: I1206 03:38:28.877875 4980 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 06 03:38:28 crc kubenswrapper[4980]: I1206 03:38:28.881994 4980 scope.go:117] "RemoveContainer" containerID="d40e0e7446f8416bf9da94fc53c01be937f9a51a11eb101157d2ddd4597720d2" Dec 06 03:38:28 crc kubenswrapper[4980]: I1206 03:38:28.900241 4980 scope.go:117] "RemoveContainer" containerID="76cff8a2d652df38b04eefc053832baf4cbb5db778ff5e525b18c2825141b118" Dec 06 03:38:28 crc kubenswrapper[4980]: I1206 03:38:28.918901 4980 scope.go:117] "RemoveContainer" containerID="206f6f19fe3197f5d6798cc86c5473bf2decb02a2860ee7ffb5efb5f60952e8c" Dec 06 03:38:28 crc kubenswrapper[4980]: I1206 03:38:28.938092 4980 scope.go:117] "RemoveContainer" containerID="c28b6895b14e73f4f3eb7a8ce93076c05092ba9b58ea0e3e879c65bcfa227a21" Dec 06 03:38:28 crc kubenswrapper[4980]: I1206 03:38:28.965791 4980 scope.go:117] "RemoveContainer" containerID="2cba89f2f1bed96d59f183628d900d9b72c91eb472e273c8bed0babcab485910" Dec 06 03:38:28 crc kubenswrapper[4980]: I1206 03:38:28.993320 4980 scope.go:117] "RemoveContainer" containerID="bafaa9670b05671fb57b658c2960bf3ee32c84e52f86a3a3bbab01d5a38ff0c4" Dec 06 03:38:28 crc kubenswrapper[4980]: E1206 03:38:28.993889 4980 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bafaa9670b05671fb57b658c2960bf3ee32c84e52f86a3a3bbab01d5a38ff0c4\": container with ID starting with bafaa9670b05671fb57b658c2960bf3ee32c84e52f86a3a3bbab01d5a38ff0c4 not found: ID does not exist" containerID="bafaa9670b05671fb57b658c2960bf3ee32c84e52f86a3a3bbab01d5a38ff0c4" Dec 06 03:38:28 crc kubenswrapper[4980]: I1206 03:38:28.993924 4980 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bafaa9670b05671fb57b658c2960bf3ee32c84e52f86a3a3bbab01d5a38ff0c4"} err="failed to get container status \"bafaa9670b05671fb57b658c2960bf3ee32c84e52f86a3a3bbab01d5a38ff0c4\": rpc error: code = NotFound desc = could not find container \"bafaa9670b05671fb57b658c2960bf3ee32c84e52f86a3a3bbab01d5a38ff0c4\": container with ID starting with bafaa9670b05671fb57b658c2960bf3ee32c84e52f86a3a3bbab01d5a38ff0c4 not found: ID does not exist" Dec 06 03:38:28 crc kubenswrapper[4980]: I1206 03:38:28.993956 4980 scope.go:117] "RemoveContainer" containerID="d40e0e7446f8416bf9da94fc53c01be937f9a51a11eb101157d2ddd4597720d2" Dec 06 03:38:28 crc kubenswrapper[4980]: E1206 03:38:28.994322 4980 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d40e0e7446f8416bf9da94fc53c01be937f9a51a11eb101157d2ddd4597720d2\": container with ID starting with d40e0e7446f8416bf9da94fc53c01be937f9a51a11eb101157d2ddd4597720d2 not found: ID does not exist" containerID="d40e0e7446f8416bf9da94fc53c01be937f9a51a11eb101157d2ddd4597720d2" Dec 06 03:38:28 crc kubenswrapper[4980]: I1206 03:38:28.994375 4980 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d40e0e7446f8416bf9da94fc53c01be937f9a51a11eb101157d2ddd4597720d2"} err="failed to get container status \"d40e0e7446f8416bf9da94fc53c01be937f9a51a11eb101157d2ddd4597720d2\": rpc error: code = NotFound desc = could not find container \"d40e0e7446f8416bf9da94fc53c01be937f9a51a11eb101157d2ddd4597720d2\": container with ID starting with d40e0e7446f8416bf9da94fc53c01be937f9a51a11eb101157d2ddd4597720d2 not found: ID does not exist" Dec 06 03:38:28 crc kubenswrapper[4980]: I1206 03:38:28.994393 4980 scope.go:117] "RemoveContainer" containerID="76cff8a2d652df38b04eefc053832baf4cbb5db778ff5e525b18c2825141b118" Dec 06 03:38:28 crc kubenswrapper[4980]: E1206 03:38:28.994827 4980 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"76cff8a2d652df38b04eefc053832baf4cbb5db778ff5e525b18c2825141b118\": container with ID starting with 76cff8a2d652df38b04eefc053832baf4cbb5db778ff5e525b18c2825141b118 not found: ID does not exist" containerID="76cff8a2d652df38b04eefc053832baf4cbb5db778ff5e525b18c2825141b118" Dec 06 03:38:28 crc kubenswrapper[4980]: I1206 03:38:28.994921 4980 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"76cff8a2d652df38b04eefc053832baf4cbb5db778ff5e525b18c2825141b118"} err="failed to get container status \"76cff8a2d652df38b04eefc053832baf4cbb5db778ff5e525b18c2825141b118\": rpc error: code = NotFound desc = could not find container \"76cff8a2d652df38b04eefc053832baf4cbb5db778ff5e525b18c2825141b118\": container with ID starting with 76cff8a2d652df38b04eefc053832baf4cbb5db778ff5e525b18c2825141b118 not found: ID does not exist" Dec 06 03:38:28 crc kubenswrapper[4980]: I1206 03:38:28.994941 4980 scope.go:117] "RemoveContainer" containerID="206f6f19fe3197f5d6798cc86c5473bf2decb02a2860ee7ffb5efb5f60952e8c" Dec 06 03:38:28 crc kubenswrapper[4980]: E1206 03:38:28.995790 4980 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"206f6f19fe3197f5d6798cc86c5473bf2decb02a2860ee7ffb5efb5f60952e8c\": container with ID starting with 206f6f19fe3197f5d6798cc86c5473bf2decb02a2860ee7ffb5efb5f60952e8c not found: ID does not exist" containerID="206f6f19fe3197f5d6798cc86c5473bf2decb02a2860ee7ffb5efb5f60952e8c" Dec 06 03:38:28 crc kubenswrapper[4980]: I1206 03:38:28.995838 4980 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"206f6f19fe3197f5d6798cc86c5473bf2decb02a2860ee7ffb5efb5f60952e8c"} err="failed to get container status \"206f6f19fe3197f5d6798cc86c5473bf2decb02a2860ee7ffb5efb5f60952e8c\": rpc error: code = NotFound desc = could not find container \"206f6f19fe3197f5d6798cc86c5473bf2decb02a2860ee7ffb5efb5f60952e8c\": container with ID starting with 206f6f19fe3197f5d6798cc86c5473bf2decb02a2860ee7ffb5efb5f60952e8c not found: ID does not exist" Dec 06 03:38:28 crc kubenswrapper[4980]: I1206 03:38:28.995868 4980 scope.go:117] "RemoveContainer" containerID="c28b6895b14e73f4f3eb7a8ce93076c05092ba9b58ea0e3e879c65bcfa227a21" Dec 06 03:38:28 crc kubenswrapper[4980]: E1206 03:38:28.996107 4980 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c28b6895b14e73f4f3eb7a8ce93076c05092ba9b58ea0e3e879c65bcfa227a21\": container with ID starting with c28b6895b14e73f4f3eb7a8ce93076c05092ba9b58ea0e3e879c65bcfa227a21 not found: ID does not exist" containerID="c28b6895b14e73f4f3eb7a8ce93076c05092ba9b58ea0e3e879c65bcfa227a21" Dec 06 03:38:28 crc kubenswrapper[4980]: I1206 03:38:28.996154 4980 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c28b6895b14e73f4f3eb7a8ce93076c05092ba9b58ea0e3e879c65bcfa227a21"} err="failed to get container status \"c28b6895b14e73f4f3eb7a8ce93076c05092ba9b58ea0e3e879c65bcfa227a21\": rpc error: code = NotFound desc = could not find container \"c28b6895b14e73f4f3eb7a8ce93076c05092ba9b58ea0e3e879c65bcfa227a21\": container with ID starting with c28b6895b14e73f4f3eb7a8ce93076c05092ba9b58ea0e3e879c65bcfa227a21 not found: ID does not exist" Dec 06 03:38:28 crc kubenswrapper[4980]: I1206 03:38:28.996172 4980 scope.go:117] "RemoveContainer" containerID="2cba89f2f1bed96d59f183628d900d9b72c91eb472e273c8bed0babcab485910" Dec 06 03:38:28 crc kubenswrapper[4980]: E1206 03:38:28.996568 4980 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2cba89f2f1bed96d59f183628d900d9b72c91eb472e273c8bed0babcab485910\": container with ID starting with 2cba89f2f1bed96d59f183628d900d9b72c91eb472e273c8bed0babcab485910 not found: ID does not exist" containerID="2cba89f2f1bed96d59f183628d900d9b72c91eb472e273c8bed0babcab485910" Dec 06 03:38:28 crc kubenswrapper[4980]: I1206 03:38:28.996612 4980 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2cba89f2f1bed96d59f183628d900d9b72c91eb472e273c8bed0babcab485910"} err="failed to get container status \"2cba89f2f1bed96d59f183628d900d9b72c91eb472e273c8bed0babcab485910\": rpc error: code = NotFound desc = could not find container \"2cba89f2f1bed96d59f183628d900d9b72c91eb472e273c8bed0babcab485910\": container with ID starting with 2cba89f2f1bed96d59f183628d900d9b72c91eb472e273c8bed0babcab485910 not found: ID does not exist" Dec 06 03:38:29 crc kubenswrapper[4980]: I1206 03:38:29.196715 4980 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f4b27818a5e8e43d0dc095d08835c792" path="/var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/volumes" Dec 06 03:38:31 crc kubenswrapper[4980]: I1206 03:38:31.187551 4980 status_manager.go:851] "Failed to get status for pod" podUID="a42aad68-42cf-4264-9c94-8678b0eea1b8" pod="openshift-authentication/oauth-openshift-558db77b4-pmmcw" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/pods/oauth-openshift-558db77b4-pmmcw\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 06 03:38:31 crc kubenswrapper[4980]: I1206 03:38:31.188390 4980 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 06 03:38:31 crc kubenswrapper[4980]: I1206 03:38:31.188843 4980 status_manager.go:851] "Failed to get status for pod" podUID="652a7feb-30e3-4e9f-b1a9-c8ef201c658c" pod="openshift-marketplace/community-operators-2vfzc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-2vfzc\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 06 03:38:31 crc kubenswrapper[4980]: I1206 03:38:31.189123 4980 status_manager.go:851] "Failed to get status for pod" podUID="43112506-9fdd-4b19-a5a5-caf864c09774" pod="openshift-marketplace/redhat-operators-m9vl6" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-m9vl6\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 06 03:38:31 crc kubenswrapper[4980]: I1206 03:38:31.189485 4980 status_manager.go:851] "Failed to get status for pod" podUID="dee63a36-cd1b-410c-b329-bd7966b1d38d" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 06 03:38:32 crc kubenswrapper[4980]: I1206 03:38:32.311953 4980 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-mkd2m" Dec 06 03:38:32 crc kubenswrapper[4980]: I1206 03:38:32.312018 4980 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-mkd2m" Dec 06 03:38:32 crc kubenswrapper[4980]: I1206 03:38:32.385829 4980 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-58d94" Dec 06 03:38:32 crc kubenswrapper[4980]: I1206 03:38:32.385897 4980 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-2vfzc" Dec 06 03:38:32 crc kubenswrapper[4980]: I1206 03:38:32.385909 4980 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-58d94" Dec 06 03:38:32 crc kubenswrapper[4980]: I1206 03:38:32.385918 4980 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-2vfzc" Dec 06 03:38:32 crc kubenswrapper[4980]: I1206 03:38:32.428375 4980 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-2vfzc" Dec 06 03:38:32 crc kubenswrapper[4980]: I1206 03:38:32.428994 4980 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-58d94" Dec 06 03:38:32 crc kubenswrapper[4980]: I1206 03:38:32.429316 4980 status_manager.go:851] "Failed to get status for pod" podUID="a42aad68-42cf-4264-9c94-8678b0eea1b8" pod="openshift-authentication/oauth-openshift-558db77b4-pmmcw" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/pods/oauth-openshift-558db77b4-pmmcw\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 06 03:38:32 crc kubenswrapper[4980]: I1206 03:38:32.429413 4980 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-mkd2m" Dec 06 03:38:32 crc kubenswrapper[4980]: I1206 03:38:32.429925 4980 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 06 03:38:32 crc kubenswrapper[4980]: I1206 03:38:32.430827 4980 status_manager.go:851] "Failed to get status for pod" podUID="43112506-9fdd-4b19-a5a5-caf864c09774" pod="openshift-marketplace/redhat-operators-m9vl6" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-m9vl6\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 06 03:38:32 crc kubenswrapper[4980]: I1206 03:38:32.431216 4980 status_manager.go:851] "Failed to get status for pod" podUID="652a7feb-30e3-4e9f-b1a9-c8ef201c658c" pod="openshift-marketplace/community-operators-2vfzc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-2vfzc\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 06 03:38:32 crc kubenswrapper[4980]: I1206 03:38:32.431591 4980 status_manager.go:851] "Failed to get status for pod" podUID="dee63a36-cd1b-410c-b329-bd7966b1d38d" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 06 03:38:32 crc kubenswrapper[4980]: I1206 03:38:32.431980 4980 status_manager.go:851] "Failed to get status for pod" podUID="a42aad68-42cf-4264-9c94-8678b0eea1b8" pod="openshift-authentication/oauth-openshift-558db77b4-pmmcw" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/pods/oauth-openshift-558db77b4-pmmcw\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 06 03:38:32 crc kubenswrapper[4980]: I1206 03:38:32.432229 4980 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 06 03:38:32 crc kubenswrapper[4980]: I1206 03:38:32.432553 4980 status_manager.go:851] "Failed to get status for pod" podUID="9c8b3374-77eb-44ad-b270-2263c6c23e28" pod="openshift-marketplace/community-operators-mkd2m" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-mkd2m\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 06 03:38:32 crc kubenswrapper[4980]: I1206 03:38:32.432869 4980 status_manager.go:851] "Failed to get status for pod" podUID="e965bb7a-1bf0-4574-aae4-2a45c75fb11c" pod="openshift-marketplace/certified-operators-58d94" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-58d94\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 06 03:38:32 crc kubenswrapper[4980]: I1206 03:38:32.433197 4980 status_manager.go:851] "Failed to get status for pod" podUID="43112506-9fdd-4b19-a5a5-caf864c09774" pod="openshift-marketplace/redhat-operators-m9vl6" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-m9vl6\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 06 03:38:32 crc kubenswrapper[4980]: I1206 03:38:32.433444 4980 status_manager.go:851] "Failed to get status for pod" podUID="652a7feb-30e3-4e9f-b1a9-c8ef201c658c" pod="openshift-marketplace/community-operators-2vfzc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-2vfzc\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 06 03:38:32 crc kubenswrapper[4980]: I1206 03:38:32.433771 4980 status_manager.go:851] "Failed to get status for pod" podUID="dee63a36-cd1b-410c-b329-bd7966b1d38d" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 06 03:38:32 crc kubenswrapper[4980]: I1206 03:38:32.917141 4980 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-kp68b" Dec 06 03:38:32 crc kubenswrapper[4980]: I1206 03:38:32.917226 4980 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-kp68b" Dec 06 03:38:32 crc kubenswrapper[4980]: I1206 03:38:32.922800 4980 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-mkd2m" Dec 06 03:38:32 crc kubenswrapper[4980]: I1206 03:38:32.924250 4980 status_manager.go:851] "Failed to get status for pod" podUID="dee63a36-cd1b-410c-b329-bd7966b1d38d" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 06 03:38:32 crc kubenswrapper[4980]: I1206 03:38:32.924869 4980 status_manager.go:851] "Failed to get status for pod" podUID="a42aad68-42cf-4264-9c94-8678b0eea1b8" pod="openshift-authentication/oauth-openshift-558db77b4-pmmcw" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/pods/oauth-openshift-558db77b4-pmmcw\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 06 03:38:32 crc kubenswrapper[4980]: I1206 03:38:32.925396 4980 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 06 03:38:32 crc kubenswrapper[4980]: I1206 03:38:32.926125 4980 status_manager.go:851] "Failed to get status for pod" podUID="9c8b3374-77eb-44ad-b270-2263c6c23e28" pod="openshift-marketplace/community-operators-mkd2m" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-mkd2m\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 06 03:38:32 crc kubenswrapper[4980]: I1206 03:38:32.926733 4980 status_manager.go:851] "Failed to get status for pod" podUID="e965bb7a-1bf0-4574-aae4-2a45c75fb11c" pod="openshift-marketplace/certified-operators-58d94" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-58d94\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 06 03:38:32 crc kubenswrapper[4980]: I1206 03:38:32.927038 4980 status_manager.go:851] "Failed to get status for pod" podUID="652a7feb-30e3-4e9f-b1a9-c8ef201c658c" pod="openshift-marketplace/community-operators-2vfzc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-2vfzc\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 06 03:38:32 crc kubenswrapper[4980]: I1206 03:38:32.927267 4980 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-2vfzc" Dec 06 03:38:32 crc kubenswrapper[4980]: I1206 03:38:32.927319 4980 status_manager.go:851] "Failed to get status for pod" podUID="43112506-9fdd-4b19-a5a5-caf864c09774" pod="openshift-marketplace/redhat-operators-m9vl6" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-m9vl6\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 06 03:38:32 crc kubenswrapper[4980]: I1206 03:38:32.927932 4980 status_manager.go:851] "Failed to get status for pod" podUID="dee63a36-cd1b-410c-b329-bd7966b1d38d" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 06 03:38:32 crc kubenswrapper[4980]: I1206 03:38:32.928392 4980 status_manager.go:851] "Failed to get status for pod" podUID="a42aad68-42cf-4264-9c94-8678b0eea1b8" pod="openshift-authentication/oauth-openshift-558db77b4-pmmcw" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/pods/oauth-openshift-558db77b4-pmmcw\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 06 03:38:32 crc kubenswrapper[4980]: I1206 03:38:32.929004 4980 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 06 03:38:32 crc kubenswrapper[4980]: I1206 03:38:32.929353 4980 status_manager.go:851] "Failed to get status for pod" podUID="9c8b3374-77eb-44ad-b270-2263c6c23e28" pod="openshift-marketplace/community-operators-mkd2m" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-mkd2m\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 06 03:38:32 crc kubenswrapper[4980]: I1206 03:38:32.931198 4980 status_manager.go:851] "Failed to get status for pod" podUID="e965bb7a-1bf0-4574-aae4-2a45c75fb11c" pod="openshift-marketplace/certified-operators-58d94" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-58d94\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 06 03:38:32 crc kubenswrapper[4980]: I1206 03:38:32.931631 4980 status_manager.go:851] "Failed to get status for pod" podUID="652a7feb-30e3-4e9f-b1a9-c8ef201c658c" pod="openshift-marketplace/community-operators-2vfzc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-2vfzc\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 06 03:38:32 crc kubenswrapper[4980]: I1206 03:38:32.932017 4980 status_manager.go:851] "Failed to get status for pod" podUID="43112506-9fdd-4b19-a5a5-caf864c09774" pod="openshift-marketplace/redhat-operators-m9vl6" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-m9vl6\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 06 03:38:32 crc kubenswrapper[4980]: I1206 03:38:32.932932 4980 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-58d94" Dec 06 03:38:32 crc kubenswrapper[4980]: I1206 03:38:32.933449 4980 status_manager.go:851] "Failed to get status for pod" podUID="a42aad68-42cf-4264-9c94-8678b0eea1b8" pod="openshift-authentication/oauth-openshift-558db77b4-pmmcw" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/pods/oauth-openshift-558db77b4-pmmcw\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 06 03:38:32 crc kubenswrapper[4980]: I1206 03:38:32.933822 4980 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 06 03:38:32 crc kubenswrapper[4980]: I1206 03:38:32.934095 4980 status_manager.go:851] "Failed to get status for pod" podUID="9c8b3374-77eb-44ad-b270-2263c6c23e28" pod="openshift-marketplace/community-operators-mkd2m" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-mkd2m\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 06 03:38:32 crc kubenswrapper[4980]: I1206 03:38:32.934425 4980 status_manager.go:851] "Failed to get status for pod" podUID="e965bb7a-1bf0-4574-aae4-2a45c75fb11c" pod="openshift-marketplace/certified-operators-58d94" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-58d94\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 06 03:38:32 crc kubenswrapper[4980]: I1206 03:38:32.934832 4980 status_manager.go:851] "Failed to get status for pod" podUID="652a7feb-30e3-4e9f-b1a9-c8ef201c658c" pod="openshift-marketplace/community-operators-2vfzc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-2vfzc\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 06 03:38:32 crc kubenswrapper[4980]: I1206 03:38:32.935046 4980 status_manager.go:851] "Failed to get status for pod" podUID="43112506-9fdd-4b19-a5a5-caf864c09774" pod="openshift-marketplace/redhat-operators-m9vl6" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-m9vl6\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 06 03:38:32 crc kubenswrapper[4980]: I1206 03:38:32.935338 4980 status_manager.go:851] "Failed to get status for pod" podUID="dee63a36-cd1b-410c-b329-bd7966b1d38d" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 06 03:38:32 crc kubenswrapper[4980]: I1206 03:38:32.968607 4980 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-kp68b" Dec 06 03:38:32 crc kubenswrapper[4980]: I1206 03:38:32.969574 4980 status_manager.go:851] "Failed to get status for pod" podUID="652a7feb-30e3-4e9f-b1a9-c8ef201c658c" pod="openshift-marketplace/community-operators-2vfzc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-2vfzc\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 06 03:38:32 crc kubenswrapper[4980]: I1206 03:38:32.969989 4980 status_manager.go:851] "Failed to get status for pod" podUID="43112506-9fdd-4b19-a5a5-caf864c09774" pod="openshift-marketplace/redhat-operators-m9vl6" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-m9vl6\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 06 03:38:32 crc kubenswrapper[4980]: I1206 03:38:32.970681 4980 status_manager.go:851] "Failed to get status for pod" podUID="dee63a36-cd1b-410c-b329-bd7966b1d38d" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 06 03:38:32 crc kubenswrapper[4980]: I1206 03:38:32.971409 4980 status_manager.go:851] "Failed to get status for pod" podUID="a42aad68-42cf-4264-9c94-8678b0eea1b8" pod="openshift-authentication/oauth-openshift-558db77b4-pmmcw" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/pods/oauth-openshift-558db77b4-pmmcw\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 06 03:38:32 crc kubenswrapper[4980]: I1206 03:38:32.971927 4980 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 06 03:38:32 crc kubenswrapper[4980]: I1206 03:38:32.972247 4980 status_manager.go:851] "Failed to get status for pod" podUID="9c8b3374-77eb-44ad-b270-2263c6c23e28" pod="openshift-marketplace/community-operators-mkd2m" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-mkd2m\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 06 03:38:32 crc kubenswrapper[4980]: I1206 03:38:32.973768 4980 status_manager.go:851] "Failed to get status for pod" podUID="e965bb7a-1bf0-4574-aae4-2a45c75fb11c" pod="openshift-marketplace/certified-operators-58d94" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-58d94\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 06 03:38:32 crc kubenswrapper[4980]: I1206 03:38:32.974215 4980 status_manager.go:851] "Failed to get status for pod" podUID="769abfb5-c18a-4cf0-a033-95b3a808867b" pod="openshift-marketplace/certified-operators-kp68b" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-kp68b\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 06 03:38:33 crc kubenswrapper[4980]: E1206 03:38:33.218486 4980 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 06 03:38:33 crc kubenswrapper[4980]: E1206 03:38:33.219359 4980 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 06 03:38:33 crc kubenswrapper[4980]: E1206 03:38:33.219735 4980 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 06 03:38:33 crc kubenswrapper[4980]: E1206 03:38:33.220094 4980 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 06 03:38:33 crc kubenswrapper[4980]: E1206 03:38:33.220382 4980 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 06 03:38:33 crc kubenswrapper[4980]: I1206 03:38:33.220425 4980 controller.go:115] "failed to update lease using latest lease, fallback to ensure lease" err="failed 5 attempts to update lease" Dec 06 03:38:33 crc kubenswrapper[4980]: E1206 03:38:33.220773 4980 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.184:6443: connect: connection refused" interval="200ms" Dec 06 03:38:33 crc kubenswrapper[4980]: E1206 03:38:33.422492 4980 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.184:6443: connect: connection refused" interval="400ms" Dec 06 03:38:33 crc kubenswrapper[4980]: E1206 03:38:33.526278 4980 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T03:38:33Z\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T03:38:33Z\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T03:38:33Z\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T03:38:33Z\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:15adb3b2133604b064893f8009a74145e4c8bb5b134d111346dcccbdd2aa9bc2\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:164fc35a19aa6cc886c8015c8ee3eba4895e76b1152cb9d795e4f3154a8533a3\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1610512706},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:544a01170a4aa6cf8322d5bffa5817113efd696e3c3e9bac6a29d2da9f9451e5\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:67f42a86b99b69b357285a6845977f967e6c825de2049c19620a78eaf99cebf3\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1222075732},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:2ebdb257c6ae11dce352087cce880750555ee33cde3006a4607a1305a2bd1b20\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:e1e1cdc371970de5a5af444ac8a3d7e6bb4086055fac0cc443b1d09c4ad585ab\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1201795919},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1b1026c62413fa239fa4ff6541fe8bda656c1281867ad6ee2c848feccb13c97e\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:2b633ebdc901d19290af4dc2d09e2b59c504c0fc15a3fba410b0ce098e2d5753\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1141987142},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792}]}}\" for node \"crc\": Patch \"https://api-int.crc.testing:6443/api/v1/nodes/crc/status?timeout=10s\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 06 03:38:33 crc kubenswrapper[4980]: E1206 03:38:33.527228 4980 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": Get \"https://api-int.crc.testing:6443/api/v1/nodes/crc?timeout=10s\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 06 03:38:33 crc kubenswrapper[4980]: E1206 03:38:33.527892 4980 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": Get \"https://api-int.crc.testing:6443/api/v1/nodes/crc?timeout=10s\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 06 03:38:33 crc kubenswrapper[4980]: E1206 03:38:33.528425 4980 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": Get \"https://api-int.crc.testing:6443/api/v1/nodes/crc?timeout=10s\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 06 03:38:33 crc kubenswrapper[4980]: E1206 03:38:33.528792 4980 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": Get \"https://api-int.crc.testing:6443/api/v1/nodes/crc?timeout=10s\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 06 03:38:33 crc kubenswrapper[4980]: E1206 03:38:33.528823 4980 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 06 03:38:33 crc kubenswrapper[4980]: E1206 03:38:33.823530 4980 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.184:6443: connect: connection refused" interval="800ms" Dec 06 03:38:33 crc kubenswrapper[4980]: I1206 03:38:33.928090 4980 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-kp68b" Dec 06 03:38:33 crc kubenswrapper[4980]: I1206 03:38:33.929007 4980 status_manager.go:851] "Failed to get status for pod" podUID="dee63a36-cd1b-410c-b329-bd7966b1d38d" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 06 03:38:33 crc kubenswrapper[4980]: I1206 03:38:33.929624 4980 status_manager.go:851] "Failed to get status for pod" podUID="a42aad68-42cf-4264-9c94-8678b0eea1b8" pod="openshift-authentication/oauth-openshift-558db77b4-pmmcw" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/pods/oauth-openshift-558db77b4-pmmcw\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 06 03:38:33 crc kubenswrapper[4980]: I1206 03:38:33.929999 4980 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 06 03:38:33 crc kubenswrapper[4980]: I1206 03:38:33.930492 4980 status_manager.go:851] "Failed to get status for pod" podUID="9c8b3374-77eb-44ad-b270-2263c6c23e28" pod="openshift-marketplace/community-operators-mkd2m" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-mkd2m\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 06 03:38:33 crc kubenswrapper[4980]: I1206 03:38:33.930944 4980 status_manager.go:851] "Failed to get status for pod" podUID="e965bb7a-1bf0-4574-aae4-2a45c75fb11c" pod="openshift-marketplace/certified-operators-58d94" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-58d94\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 06 03:38:33 crc kubenswrapper[4980]: I1206 03:38:33.931304 4980 status_manager.go:851] "Failed to get status for pod" podUID="769abfb5-c18a-4cf0-a033-95b3a808867b" pod="openshift-marketplace/certified-operators-kp68b" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-kp68b\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 06 03:38:33 crc kubenswrapper[4980]: I1206 03:38:33.932504 4980 status_manager.go:851] "Failed to get status for pod" podUID="652a7feb-30e3-4e9f-b1a9-c8ef201c658c" pod="openshift-marketplace/community-operators-2vfzc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-2vfzc\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 06 03:38:33 crc kubenswrapper[4980]: I1206 03:38:33.932947 4980 status_manager.go:851] "Failed to get status for pod" podUID="43112506-9fdd-4b19-a5a5-caf864c09774" pod="openshift-marketplace/redhat-operators-m9vl6" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-m9vl6\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 06 03:38:34 crc kubenswrapper[4980]: E1206 03:38:34.112057 4980 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/events\": dial tcp 38.102.83.184:6443: connect: connection refused" event="&Event{ObjectMeta:{community-operators-2vfzc.187e8326997e5a92 openshift-marketplace 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-marketplace,Name:community-operators-2vfzc,UID:652a7feb-30e3-4e9f-b1a9-c8ef201c658c,APIVersion:v1,ResourceVersion:28261,FieldPath:spec.containers{registry-server},},Reason:Pulling,Message:Pulling image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\",Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-06 03:38:25.783003794 +0000 UTC m=+265.027755065,LastTimestamp:2025-12-06 03:38:25.783003794 +0000 UTC m=+265.027755065,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Dec 06 03:38:34 crc kubenswrapper[4980]: I1206 03:38:34.216198 4980 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-gxcqb" Dec 06 03:38:34 crc kubenswrapper[4980]: I1206 03:38:34.216286 4980 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-gxcqb" Dec 06 03:38:34 crc kubenswrapper[4980]: I1206 03:38:34.262775 4980 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-gxcqb" Dec 06 03:38:34 crc kubenswrapper[4980]: I1206 03:38:34.263561 4980 status_manager.go:851] "Failed to get status for pod" podUID="dee63a36-cd1b-410c-b329-bd7966b1d38d" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 06 03:38:34 crc kubenswrapper[4980]: I1206 03:38:34.264075 4980 status_manager.go:851] "Failed to get status for pod" podUID="a42aad68-42cf-4264-9c94-8678b0eea1b8" pod="openshift-authentication/oauth-openshift-558db77b4-pmmcw" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/pods/oauth-openshift-558db77b4-pmmcw\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 06 03:38:34 crc kubenswrapper[4980]: I1206 03:38:34.264418 4980 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 06 03:38:34 crc kubenswrapper[4980]: I1206 03:38:34.264693 4980 status_manager.go:851] "Failed to get status for pod" podUID="9c8b3374-77eb-44ad-b270-2263c6c23e28" pod="openshift-marketplace/community-operators-mkd2m" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-mkd2m\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 06 03:38:34 crc kubenswrapper[4980]: I1206 03:38:34.265002 4980 status_manager.go:851] "Failed to get status for pod" podUID="e965bb7a-1bf0-4574-aae4-2a45c75fb11c" pod="openshift-marketplace/certified-operators-58d94" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-58d94\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 06 03:38:34 crc kubenswrapper[4980]: I1206 03:38:34.265362 4980 status_manager.go:851] "Failed to get status for pod" podUID="769abfb5-c18a-4cf0-a033-95b3a808867b" pod="openshift-marketplace/certified-operators-kp68b" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-kp68b\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 06 03:38:34 crc kubenswrapper[4980]: I1206 03:38:34.265667 4980 status_manager.go:851] "Failed to get status for pod" podUID="652a7feb-30e3-4e9f-b1a9-c8ef201c658c" pod="openshift-marketplace/community-operators-2vfzc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-2vfzc\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 06 03:38:34 crc kubenswrapper[4980]: I1206 03:38:34.266018 4980 status_manager.go:851] "Failed to get status for pod" podUID="2ce9bc16-3d2a-4feb-82c1-0f42518196cf" pod="openshift-marketplace/redhat-marketplace-gxcqb" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-gxcqb\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 06 03:38:34 crc kubenswrapper[4980]: I1206 03:38:34.266451 4980 status_manager.go:851] "Failed to get status for pod" podUID="43112506-9fdd-4b19-a5a5-caf864c09774" pod="openshift-marketplace/redhat-operators-m9vl6" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-m9vl6\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 06 03:38:34 crc kubenswrapper[4980]: I1206 03:38:34.592536 4980 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-xhnw8" Dec 06 03:38:34 crc kubenswrapper[4980]: I1206 03:38:34.592605 4980 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-xhnw8" Dec 06 03:38:34 crc kubenswrapper[4980]: E1206 03:38:34.624854 4980 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.184:6443: connect: connection refused" interval="1.6s" Dec 06 03:38:34 crc kubenswrapper[4980]: I1206 03:38:34.650936 4980 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-xhnw8" Dec 06 03:38:34 crc kubenswrapper[4980]: I1206 03:38:34.651373 4980 status_manager.go:851] "Failed to get status for pod" podUID="3caa51b4-a311-4498-ac9b-10c20989684a" pod="openshift-marketplace/redhat-marketplace-xhnw8" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-xhnw8\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 06 03:38:34 crc kubenswrapper[4980]: I1206 03:38:34.651894 4980 status_manager.go:851] "Failed to get status for pod" podUID="dee63a36-cd1b-410c-b329-bd7966b1d38d" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 06 03:38:34 crc kubenswrapper[4980]: I1206 03:38:34.652401 4980 status_manager.go:851] "Failed to get status for pod" podUID="a42aad68-42cf-4264-9c94-8678b0eea1b8" pod="openshift-authentication/oauth-openshift-558db77b4-pmmcw" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/pods/oauth-openshift-558db77b4-pmmcw\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 06 03:38:34 crc kubenswrapper[4980]: I1206 03:38:34.652651 4980 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 06 03:38:34 crc kubenswrapper[4980]: I1206 03:38:34.653057 4980 status_manager.go:851] "Failed to get status for pod" podUID="9c8b3374-77eb-44ad-b270-2263c6c23e28" pod="openshift-marketplace/community-operators-mkd2m" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-mkd2m\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 06 03:38:34 crc kubenswrapper[4980]: I1206 03:38:34.653440 4980 status_manager.go:851] "Failed to get status for pod" podUID="e965bb7a-1bf0-4574-aae4-2a45c75fb11c" pod="openshift-marketplace/certified-operators-58d94" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-58d94\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 06 03:38:34 crc kubenswrapper[4980]: I1206 03:38:34.653679 4980 status_manager.go:851] "Failed to get status for pod" podUID="769abfb5-c18a-4cf0-a033-95b3a808867b" pod="openshift-marketplace/certified-operators-kp68b" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-kp68b\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 06 03:38:34 crc kubenswrapper[4980]: I1206 03:38:34.653919 4980 status_manager.go:851] "Failed to get status for pod" podUID="43112506-9fdd-4b19-a5a5-caf864c09774" pod="openshift-marketplace/redhat-operators-m9vl6" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-m9vl6\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 06 03:38:34 crc kubenswrapper[4980]: I1206 03:38:34.654288 4980 status_manager.go:851] "Failed to get status for pod" podUID="652a7feb-30e3-4e9f-b1a9-c8ef201c658c" pod="openshift-marketplace/community-operators-2vfzc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-2vfzc\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 06 03:38:34 crc kubenswrapper[4980]: I1206 03:38:34.654777 4980 status_manager.go:851] "Failed to get status for pod" podUID="2ce9bc16-3d2a-4feb-82c1-0f42518196cf" pod="openshift-marketplace/redhat-marketplace-gxcqb" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-gxcqb\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 06 03:38:34 crc kubenswrapper[4980]: I1206 03:38:34.936014 4980 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-xhnw8" Dec 06 03:38:34 crc kubenswrapper[4980]: I1206 03:38:34.936882 4980 status_manager.go:851] "Failed to get status for pod" podUID="2ce9bc16-3d2a-4feb-82c1-0f42518196cf" pod="openshift-marketplace/redhat-marketplace-gxcqb" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-gxcqb\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 06 03:38:34 crc kubenswrapper[4980]: I1206 03:38:34.937586 4980 status_manager.go:851] "Failed to get status for pod" podUID="43112506-9fdd-4b19-a5a5-caf864c09774" pod="openshift-marketplace/redhat-operators-m9vl6" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-m9vl6\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 06 03:38:34 crc kubenswrapper[4980]: I1206 03:38:34.938259 4980 status_manager.go:851] "Failed to get status for pod" podUID="652a7feb-30e3-4e9f-b1a9-c8ef201c658c" pod="openshift-marketplace/community-operators-2vfzc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-2vfzc\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 06 03:38:34 crc kubenswrapper[4980]: I1206 03:38:34.938821 4980 status_manager.go:851] "Failed to get status for pod" podUID="3caa51b4-a311-4498-ac9b-10c20989684a" pod="openshift-marketplace/redhat-marketplace-xhnw8" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-xhnw8\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 06 03:38:34 crc kubenswrapper[4980]: I1206 03:38:34.939297 4980 status_manager.go:851] "Failed to get status for pod" podUID="dee63a36-cd1b-410c-b329-bd7966b1d38d" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 06 03:38:34 crc kubenswrapper[4980]: I1206 03:38:34.939784 4980 status_manager.go:851] "Failed to get status for pod" podUID="a42aad68-42cf-4264-9c94-8678b0eea1b8" pod="openshift-authentication/oauth-openshift-558db77b4-pmmcw" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/pods/oauth-openshift-558db77b4-pmmcw\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 06 03:38:34 crc kubenswrapper[4980]: I1206 03:38:34.940316 4980 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 06 03:38:34 crc kubenswrapper[4980]: I1206 03:38:34.940814 4980 status_manager.go:851] "Failed to get status for pod" podUID="9c8b3374-77eb-44ad-b270-2263c6c23e28" pod="openshift-marketplace/community-operators-mkd2m" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-mkd2m\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 06 03:38:34 crc kubenswrapper[4980]: I1206 03:38:34.941212 4980 status_manager.go:851] "Failed to get status for pod" podUID="e965bb7a-1bf0-4574-aae4-2a45c75fb11c" pod="openshift-marketplace/certified-operators-58d94" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-58d94\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 06 03:38:34 crc kubenswrapper[4980]: I1206 03:38:34.941587 4980 status_manager.go:851] "Failed to get status for pod" podUID="769abfb5-c18a-4cf0-a033-95b3a808867b" pod="openshift-marketplace/certified-operators-kp68b" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-kp68b\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 06 03:38:34 crc kubenswrapper[4980]: I1206 03:38:34.962695 4980 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-gxcqb" Dec 06 03:38:34 crc kubenswrapper[4980]: I1206 03:38:34.963619 4980 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 06 03:38:34 crc kubenswrapper[4980]: I1206 03:38:34.964359 4980 status_manager.go:851] "Failed to get status for pod" podUID="9c8b3374-77eb-44ad-b270-2263c6c23e28" pod="openshift-marketplace/community-operators-mkd2m" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-mkd2m\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 06 03:38:34 crc kubenswrapper[4980]: I1206 03:38:34.964903 4980 status_manager.go:851] "Failed to get status for pod" podUID="e965bb7a-1bf0-4574-aae4-2a45c75fb11c" pod="openshift-marketplace/certified-operators-58d94" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-58d94\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 06 03:38:34 crc kubenswrapper[4980]: I1206 03:38:34.965352 4980 status_manager.go:851] "Failed to get status for pod" podUID="769abfb5-c18a-4cf0-a033-95b3a808867b" pod="openshift-marketplace/certified-operators-kp68b" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-kp68b\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 06 03:38:34 crc kubenswrapper[4980]: I1206 03:38:34.965928 4980 status_manager.go:851] "Failed to get status for pod" podUID="652a7feb-30e3-4e9f-b1a9-c8ef201c658c" pod="openshift-marketplace/community-operators-2vfzc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-2vfzc\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 06 03:38:34 crc kubenswrapper[4980]: I1206 03:38:34.966339 4980 status_manager.go:851] "Failed to get status for pod" podUID="2ce9bc16-3d2a-4feb-82c1-0f42518196cf" pod="openshift-marketplace/redhat-marketplace-gxcqb" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-gxcqb\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 06 03:38:34 crc kubenswrapper[4980]: I1206 03:38:34.966862 4980 status_manager.go:851] "Failed to get status for pod" podUID="43112506-9fdd-4b19-a5a5-caf864c09774" pod="openshift-marketplace/redhat-operators-m9vl6" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-m9vl6\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 06 03:38:34 crc kubenswrapper[4980]: I1206 03:38:34.967453 4980 status_manager.go:851] "Failed to get status for pod" podUID="3caa51b4-a311-4498-ac9b-10c20989684a" pod="openshift-marketplace/redhat-marketplace-xhnw8" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-xhnw8\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 06 03:38:34 crc kubenswrapper[4980]: I1206 03:38:34.967863 4980 status_manager.go:851] "Failed to get status for pod" podUID="dee63a36-cd1b-410c-b329-bd7966b1d38d" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 06 03:38:34 crc kubenswrapper[4980]: I1206 03:38:34.968280 4980 status_manager.go:851] "Failed to get status for pod" podUID="a42aad68-42cf-4264-9c94-8678b0eea1b8" pod="openshift-authentication/oauth-openshift-558db77b4-pmmcw" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/pods/oauth-openshift-558db77b4-pmmcw\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 06 03:38:35 crc kubenswrapper[4980]: I1206 03:38:35.267651 4980 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-m9vl6" Dec 06 03:38:35 crc kubenswrapper[4980]: I1206 03:38:35.268405 4980 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-m9vl6" Dec 06 03:38:35 crc kubenswrapper[4980]: I1206 03:38:35.306915 4980 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-m9vl6" Dec 06 03:38:35 crc kubenswrapper[4980]: I1206 03:38:35.307819 4980 status_manager.go:851] "Failed to get status for pod" podUID="a42aad68-42cf-4264-9c94-8678b0eea1b8" pod="openshift-authentication/oauth-openshift-558db77b4-pmmcw" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/pods/oauth-openshift-558db77b4-pmmcw\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 06 03:38:35 crc kubenswrapper[4980]: I1206 03:38:35.308474 4980 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 06 03:38:35 crc kubenswrapper[4980]: I1206 03:38:35.308951 4980 status_manager.go:851] "Failed to get status for pod" podUID="9c8b3374-77eb-44ad-b270-2263c6c23e28" pod="openshift-marketplace/community-operators-mkd2m" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-mkd2m\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 06 03:38:35 crc kubenswrapper[4980]: I1206 03:38:35.309230 4980 status_manager.go:851] "Failed to get status for pod" podUID="e965bb7a-1bf0-4574-aae4-2a45c75fb11c" pod="openshift-marketplace/certified-operators-58d94" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-58d94\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 06 03:38:35 crc kubenswrapper[4980]: I1206 03:38:35.309833 4980 status_manager.go:851] "Failed to get status for pod" podUID="769abfb5-c18a-4cf0-a033-95b3a808867b" pod="openshift-marketplace/certified-operators-kp68b" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-kp68b\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 06 03:38:35 crc kubenswrapper[4980]: I1206 03:38:35.310090 4980 status_manager.go:851] "Failed to get status for pod" podUID="652a7feb-30e3-4e9f-b1a9-c8ef201c658c" pod="openshift-marketplace/community-operators-2vfzc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-2vfzc\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 06 03:38:35 crc kubenswrapper[4980]: I1206 03:38:35.310471 4980 status_manager.go:851] "Failed to get status for pod" podUID="2ce9bc16-3d2a-4feb-82c1-0f42518196cf" pod="openshift-marketplace/redhat-marketplace-gxcqb" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-gxcqb\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 06 03:38:35 crc kubenswrapper[4980]: I1206 03:38:35.310719 4980 status_manager.go:851] "Failed to get status for pod" podUID="43112506-9fdd-4b19-a5a5-caf864c09774" pod="openshift-marketplace/redhat-operators-m9vl6" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-m9vl6\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 06 03:38:35 crc kubenswrapper[4980]: I1206 03:38:35.311114 4980 status_manager.go:851] "Failed to get status for pod" podUID="3caa51b4-a311-4498-ac9b-10c20989684a" pod="openshift-marketplace/redhat-marketplace-xhnw8" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-xhnw8\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 06 03:38:35 crc kubenswrapper[4980]: I1206 03:38:35.311995 4980 status_manager.go:851] "Failed to get status for pod" podUID="dee63a36-cd1b-410c-b329-bd7966b1d38d" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 06 03:38:35 crc kubenswrapper[4980]: I1206 03:38:35.636221 4980 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-bl7hx" Dec 06 03:38:35 crc kubenswrapper[4980]: I1206 03:38:35.636942 4980 status_manager.go:851] "Failed to get status for pod" podUID="a42aad68-42cf-4264-9c94-8678b0eea1b8" pod="openshift-authentication/oauth-openshift-558db77b4-pmmcw" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/pods/oauth-openshift-558db77b4-pmmcw\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 06 03:38:35 crc kubenswrapper[4980]: I1206 03:38:35.637137 4980 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 06 03:38:35 crc kubenswrapper[4980]: I1206 03:38:35.637279 4980 status_manager.go:851] "Failed to get status for pod" podUID="9c8b3374-77eb-44ad-b270-2263c6c23e28" pod="openshift-marketplace/community-operators-mkd2m" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-mkd2m\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 06 03:38:35 crc kubenswrapper[4980]: I1206 03:38:35.637408 4980 status_manager.go:851] "Failed to get status for pod" podUID="e965bb7a-1bf0-4574-aae4-2a45c75fb11c" pod="openshift-marketplace/certified-operators-58d94" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-58d94\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 06 03:38:35 crc kubenswrapper[4980]: I1206 03:38:35.637591 4980 status_manager.go:851] "Failed to get status for pod" podUID="769abfb5-c18a-4cf0-a033-95b3a808867b" pod="openshift-marketplace/certified-operators-kp68b" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-kp68b\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 06 03:38:35 crc kubenswrapper[4980]: I1206 03:38:35.637748 4980 status_manager.go:851] "Failed to get status for pod" podUID="56a102c7-f341-4076-89dc-44428e77c164" pod="openshift-marketplace/redhat-operators-bl7hx" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-bl7hx\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 06 03:38:35 crc kubenswrapper[4980]: I1206 03:38:35.637920 4980 status_manager.go:851] "Failed to get status for pod" podUID="652a7feb-30e3-4e9f-b1a9-c8ef201c658c" pod="openshift-marketplace/community-operators-2vfzc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-2vfzc\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 06 03:38:35 crc kubenswrapper[4980]: I1206 03:38:35.638100 4980 status_manager.go:851] "Failed to get status for pod" podUID="2ce9bc16-3d2a-4feb-82c1-0f42518196cf" pod="openshift-marketplace/redhat-marketplace-gxcqb" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-gxcqb\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 06 03:38:35 crc kubenswrapper[4980]: I1206 03:38:35.638393 4980 status_manager.go:851] "Failed to get status for pod" podUID="43112506-9fdd-4b19-a5a5-caf864c09774" pod="openshift-marketplace/redhat-operators-m9vl6" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-m9vl6\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 06 03:38:35 crc kubenswrapper[4980]: I1206 03:38:35.638860 4980 status_manager.go:851] "Failed to get status for pod" podUID="3caa51b4-a311-4498-ac9b-10c20989684a" pod="openshift-marketplace/redhat-marketplace-xhnw8" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-xhnw8\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 06 03:38:35 crc kubenswrapper[4980]: I1206 03:38:35.639045 4980 status_manager.go:851] "Failed to get status for pod" podUID="dee63a36-cd1b-410c-b329-bd7966b1d38d" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 06 03:38:35 crc kubenswrapper[4980]: I1206 03:38:35.673285 4980 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-bl7hx" Dec 06 03:38:35 crc kubenswrapper[4980]: I1206 03:38:35.673764 4980 status_manager.go:851] "Failed to get status for pod" podUID="dee63a36-cd1b-410c-b329-bd7966b1d38d" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 06 03:38:35 crc kubenswrapper[4980]: I1206 03:38:35.674259 4980 status_manager.go:851] "Failed to get status for pod" podUID="a42aad68-42cf-4264-9c94-8678b0eea1b8" pod="openshift-authentication/oauth-openshift-558db77b4-pmmcw" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/pods/oauth-openshift-558db77b4-pmmcw\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 06 03:38:35 crc kubenswrapper[4980]: I1206 03:38:35.674844 4980 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 06 03:38:35 crc kubenswrapper[4980]: I1206 03:38:35.675077 4980 status_manager.go:851] "Failed to get status for pod" podUID="9c8b3374-77eb-44ad-b270-2263c6c23e28" pod="openshift-marketplace/community-operators-mkd2m" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-mkd2m\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 06 03:38:35 crc kubenswrapper[4980]: I1206 03:38:35.675801 4980 status_manager.go:851] "Failed to get status for pod" podUID="e965bb7a-1bf0-4574-aae4-2a45c75fb11c" pod="openshift-marketplace/certified-operators-58d94" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-58d94\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 06 03:38:35 crc kubenswrapper[4980]: I1206 03:38:35.676117 4980 status_manager.go:851] "Failed to get status for pod" podUID="769abfb5-c18a-4cf0-a033-95b3a808867b" pod="openshift-marketplace/certified-operators-kp68b" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-kp68b\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 06 03:38:35 crc kubenswrapper[4980]: I1206 03:38:35.676335 4980 status_manager.go:851] "Failed to get status for pod" podUID="56a102c7-f341-4076-89dc-44428e77c164" pod="openshift-marketplace/redhat-operators-bl7hx" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-bl7hx\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 06 03:38:35 crc kubenswrapper[4980]: I1206 03:38:35.676567 4980 status_manager.go:851] "Failed to get status for pod" podUID="652a7feb-30e3-4e9f-b1a9-c8ef201c658c" pod="openshift-marketplace/community-operators-2vfzc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-2vfzc\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 06 03:38:35 crc kubenswrapper[4980]: I1206 03:38:35.676761 4980 status_manager.go:851] "Failed to get status for pod" podUID="2ce9bc16-3d2a-4feb-82c1-0f42518196cf" pod="openshift-marketplace/redhat-marketplace-gxcqb" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-gxcqb\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 06 03:38:35 crc kubenswrapper[4980]: I1206 03:38:35.676982 4980 status_manager.go:851] "Failed to get status for pod" podUID="43112506-9fdd-4b19-a5a5-caf864c09774" pod="openshift-marketplace/redhat-operators-m9vl6" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-m9vl6\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 06 03:38:35 crc kubenswrapper[4980]: I1206 03:38:35.677412 4980 status_manager.go:851] "Failed to get status for pod" podUID="3caa51b4-a311-4498-ac9b-10c20989684a" pod="openshift-marketplace/redhat-marketplace-xhnw8" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-xhnw8\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 06 03:38:35 crc kubenswrapper[4980]: I1206 03:38:35.969601 4980 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-m9vl6" Dec 06 03:38:35 crc kubenswrapper[4980]: I1206 03:38:35.970399 4980 status_manager.go:851] "Failed to get status for pod" podUID="a42aad68-42cf-4264-9c94-8678b0eea1b8" pod="openshift-authentication/oauth-openshift-558db77b4-pmmcw" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/pods/oauth-openshift-558db77b4-pmmcw\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 06 03:38:35 crc kubenswrapper[4980]: I1206 03:38:35.971102 4980 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 06 03:38:35 crc kubenswrapper[4980]: I1206 03:38:35.971583 4980 status_manager.go:851] "Failed to get status for pod" podUID="9c8b3374-77eb-44ad-b270-2263c6c23e28" pod="openshift-marketplace/community-operators-mkd2m" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-mkd2m\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 06 03:38:35 crc kubenswrapper[4980]: I1206 03:38:35.971912 4980 status_manager.go:851] "Failed to get status for pod" podUID="e965bb7a-1bf0-4574-aae4-2a45c75fb11c" pod="openshift-marketplace/certified-operators-58d94" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-58d94\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 06 03:38:35 crc kubenswrapper[4980]: I1206 03:38:35.972224 4980 status_manager.go:851] "Failed to get status for pod" podUID="56a102c7-f341-4076-89dc-44428e77c164" pod="openshift-marketplace/redhat-operators-bl7hx" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-bl7hx\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 06 03:38:35 crc kubenswrapper[4980]: I1206 03:38:35.972582 4980 status_manager.go:851] "Failed to get status for pod" podUID="769abfb5-c18a-4cf0-a033-95b3a808867b" pod="openshift-marketplace/certified-operators-kp68b" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-kp68b\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 06 03:38:35 crc kubenswrapper[4980]: I1206 03:38:35.972953 4980 status_manager.go:851] "Failed to get status for pod" podUID="2ce9bc16-3d2a-4feb-82c1-0f42518196cf" pod="openshift-marketplace/redhat-marketplace-gxcqb" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-gxcqb\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 06 03:38:35 crc kubenswrapper[4980]: I1206 03:38:35.973318 4980 status_manager.go:851] "Failed to get status for pod" podUID="43112506-9fdd-4b19-a5a5-caf864c09774" pod="openshift-marketplace/redhat-operators-m9vl6" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-m9vl6\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 06 03:38:35 crc kubenswrapper[4980]: I1206 03:38:35.973680 4980 status_manager.go:851] "Failed to get status for pod" podUID="652a7feb-30e3-4e9f-b1a9-c8ef201c658c" pod="openshift-marketplace/community-operators-2vfzc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-2vfzc\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 06 03:38:35 crc kubenswrapper[4980]: I1206 03:38:35.974014 4980 status_manager.go:851] "Failed to get status for pod" podUID="3caa51b4-a311-4498-ac9b-10c20989684a" pod="openshift-marketplace/redhat-marketplace-xhnw8" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-xhnw8\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 06 03:38:35 crc kubenswrapper[4980]: I1206 03:38:35.974291 4980 status_manager.go:851] "Failed to get status for pod" podUID="dee63a36-cd1b-410c-b329-bd7966b1d38d" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 06 03:38:36 crc kubenswrapper[4980]: E1206 03:38:36.225797 4980 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.184:6443: connect: connection refused" interval="3.2s" Dec 06 03:38:38 crc kubenswrapper[4980]: I1206 03:38:38.184208 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 06 03:38:38 crc kubenswrapper[4980]: I1206 03:38:38.185187 4980 status_manager.go:851] "Failed to get status for pod" podUID="a42aad68-42cf-4264-9c94-8678b0eea1b8" pod="openshift-authentication/oauth-openshift-558db77b4-pmmcw" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/pods/oauth-openshift-558db77b4-pmmcw\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 06 03:38:38 crc kubenswrapper[4980]: I1206 03:38:38.185833 4980 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 06 03:38:38 crc kubenswrapper[4980]: I1206 03:38:38.186262 4980 status_manager.go:851] "Failed to get status for pod" podUID="e965bb7a-1bf0-4574-aae4-2a45c75fb11c" pod="openshift-marketplace/certified-operators-58d94" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-58d94\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 06 03:38:38 crc kubenswrapper[4980]: I1206 03:38:38.186586 4980 status_manager.go:851] "Failed to get status for pod" podUID="9c8b3374-77eb-44ad-b270-2263c6c23e28" pod="openshift-marketplace/community-operators-mkd2m" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-mkd2m\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 06 03:38:38 crc kubenswrapper[4980]: I1206 03:38:38.186824 4980 status_manager.go:851] "Failed to get status for pod" podUID="769abfb5-c18a-4cf0-a033-95b3a808867b" pod="openshift-marketplace/certified-operators-kp68b" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-kp68b\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 06 03:38:38 crc kubenswrapper[4980]: I1206 03:38:38.187076 4980 status_manager.go:851] "Failed to get status for pod" podUID="56a102c7-f341-4076-89dc-44428e77c164" pod="openshift-marketplace/redhat-operators-bl7hx" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-bl7hx\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 06 03:38:38 crc kubenswrapper[4980]: I1206 03:38:38.187394 4980 status_manager.go:851] "Failed to get status for pod" podUID="652a7feb-30e3-4e9f-b1a9-c8ef201c658c" pod="openshift-marketplace/community-operators-2vfzc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-2vfzc\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 06 03:38:38 crc kubenswrapper[4980]: I1206 03:38:38.187786 4980 status_manager.go:851] "Failed to get status for pod" podUID="2ce9bc16-3d2a-4feb-82c1-0f42518196cf" pod="openshift-marketplace/redhat-marketplace-gxcqb" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-gxcqb\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 06 03:38:38 crc kubenswrapper[4980]: I1206 03:38:38.188058 4980 status_manager.go:851] "Failed to get status for pod" podUID="43112506-9fdd-4b19-a5a5-caf864c09774" pod="openshift-marketplace/redhat-operators-m9vl6" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-m9vl6\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 06 03:38:38 crc kubenswrapper[4980]: I1206 03:38:38.188383 4980 status_manager.go:851] "Failed to get status for pod" podUID="3caa51b4-a311-4498-ac9b-10c20989684a" pod="openshift-marketplace/redhat-marketplace-xhnw8" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-xhnw8\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 06 03:38:38 crc kubenswrapper[4980]: I1206 03:38:38.188738 4980 status_manager.go:851] "Failed to get status for pod" podUID="dee63a36-cd1b-410c-b329-bd7966b1d38d" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 06 03:38:38 crc kubenswrapper[4980]: I1206 03:38:38.200654 4980 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="7f958150-549d-4c80-8a7e-23c39e4f9500" Dec 06 03:38:38 crc kubenswrapper[4980]: I1206 03:38:38.200691 4980 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="7f958150-549d-4c80-8a7e-23c39e4f9500" Dec 06 03:38:38 crc kubenswrapper[4980]: E1206 03:38:38.201223 4980 mirror_client.go:138] "Failed deleting a mirror pod" err="Delete \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.184:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 06 03:38:38 crc kubenswrapper[4980]: I1206 03:38:38.201839 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 06 03:38:38 crc kubenswrapper[4980]: W1206 03:38:38.228992 4980 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod71bb4a3aecc4ba5b26c4b7318770ce13.slice/crio-788864b97b9196b478454c9165e5135aa5024d7ab2872e0d0c145ae74b8509d0 WatchSource:0}: Error finding container 788864b97b9196b478454c9165e5135aa5024d7ab2872e0d0c145ae74b8509d0: Status 404 returned error can't find the container with id 788864b97b9196b478454c9165e5135aa5024d7ab2872e0d0c145ae74b8509d0 Dec 06 03:38:38 crc kubenswrapper[4980]: I1206 03:38:38.918863 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"788864b97b9196b478454c9165e5135aa5024d7ab2872e0d0c145ae74b8509d0"} Dec 06 03:38:39 crc kubenswrapper[4980]: E1206 03:38:39.427322 4980 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.184:6443: connect: connection refused" interval="6.4s" Dec 06 03:38:39 crc kubenswrapper[4980]: I1206 03:38:39.927813 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"61e3b27c508dae7d50adf47d5bf73a54f4900e0c7a28d6654e59bd9f37c3162f"} Dec 06 03:38:39 crc kubenswrapper[4980]: I1206 03:38:39.928105 4980 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="7f958150-549d-4c80-8a7e-23c39e4f9500" Dec 06 03:38:39 crc kubenswrapper[4980]: I1206 03:38:39.928123 4980 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="7f958150-549d-4c80-8a7e-23c39e4f9500" Dec 06 03:38:39 crc kubenswrapper[4980]: E1206 03:38:39.928646 4980 mirror_client.go:138] "Failed deleting a mirror pod" err="Delete \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.184:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 06 03:38:39 crc kubenswrapper[4980]: I1206 03:38:39.928675 4980 status_manager.go:851] "Failed to get status for pod" podUID="a42aad68-42cf-4264-9c94-8678b0eea1b8" pod="openshift-authentication/oauth-openshift-558db77b4-pmmcw" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/pods/oauth-openshift-558db77b4-pmmcw\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 06 03:38:39 crc kubenswrapper[4980]: I1206 03:38:39.929355 4980 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 06 03:38:39 crc kubenswrapper[4980]: I1206 03:38:39.929948 4980 status_manager.go:851] "Failed to get status for pod" podUID="9c8b3374-77eb-44ad-b270-2263c6c23e28" pod="openshift-marketplace/community-operators-mkd2m" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-mkd2m\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 06 03:38:39 crc kubenswrapper[4980]: I1206 03:38:39.930476 4980 status_manager.go:851] "Failed to get status for pod" podUID="e965bb7a-1bf0-4574-aae4-2a45c75fb11c" pod="openshift-marketplace/certified-operators-58d94" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-58d94\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 06 03:38:39 crc kubenswrapper[4980]: I1206 03:38:39.930974 4980 status_manager.go:851] "Failed to get status for pod" podUID="769abfb5-c18a-4cf0-a033-95b3a808867b" pod="openshift-marketplace/certified-operators-kp68b" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-kp68b\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 06 03:38:39 crc kubenswrapper[4980]: I1206 03:38:39.931344 4980 status_manager.go:851] "Failed to get status for pod" podUID="56a102c7-f341-4076-89dc-44428e77c164" pod="openshift-marketplace/redhat-operators-bl7hx" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-bl7hx\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 06 03:38:39 crc kubenswrapper[4980]: I1206 03:38:39.931878 4980 status_manager.go:851] "Failed to get status for pod" podUID="652a7feb-30e3-4e9f-b1a9-c8ef201c658c" pod="openshift-marketplace/community-operators-2vfzc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-2vfzc\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 06 03:38:39 crc kubenswrapper[4980]: I1206 03:38:39.932409 4980 status_manager.go:851] "Failed to get status for pod" podUID="2ce9bc16-3d2a-4feb-82c1-0f42518196cf" pod="openshift-marketplace/redhat-marketplace-gxcqb" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-gxcqb\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 06 03:38:39 crc kubenswrapper[4980]: I1206 03:38:39.933234 4980 status_manager.go:851] "Failed to get status for pod" podUID="43112506-9fdd-4b19-a5a5-caf864c09774" pod="openshift-marketplace/redhat-operators-m9vl6" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-m9vl6\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 06 03:38:39 crc kubenswrapper[4980]: I1206 03:38:39.933716 4980 status_manager.go:851] "Failed to get status for pod" podUID="3caa51b4-a311-4498-ac9b-10c20989684a" pod="openshift-marketplace/redhat-marketplace-xhnw8" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-xhnw8\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 06 03:38:39 crc kubenswrapper[4980]: I1206 03:38:39.934242 4980 status_manager.go:851] "Failed to get status for pod" podUID="dee63a36-cd1b-410c-b329-bd7966b1d38d" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 06 03:38:40 crc kubenswrapper[4980]: I1206 03:38:40.937101 4980 generic.go:334] "Generic (PLEG): container finished" podID="71bb4a3aecc4ba5b26c4b7318770ce13" containerID="61e3b27c508dae7d50adf47d5bf73a54f4900e0c7a28d6654e59bd9f37c3162f" exitCode=0 Dec 06 03:38:40 crc kubenswrapper[4980]: I1206 03:38:40.937151 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerDied","Data":"61e3b27c508dae7d50adf47d5bf73a54f4900e0c7a28d6654e59bd9f37c3162f"} Dec 06 03:38:40 crc kubenswrapper[4980]: I1206 03:38:40.937353 4980 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="7f958150-549d-4c80-8a7e-23c39e4f9500" Dec 06 03:38:40 crc kubenswrapper[4980]: I1206 03:38:40.937378 4980 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="7f958150-549d-4c80-8a7e-23c39e4f9500" Dec 06 03:38:40 crc kubenswrapper[4980]: E1206 03:38:40.938869 4980 mirror_client.go:138] "Failed deleting a mirror pod" err="Delete \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.184:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 06 03:38:40 crc kubenswrapper[4980]: I1206 03:38:40.939164 4980 status_manager.go:851] "Failed to get status for pod" podUID="dee63a36-cd1b-410c-b329-bd7966b1d38d" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 06 03:38:40 crc kubenswrapper[4980]: I1206 03:38:40.940856 4980 status_manager.go:851] "Failed to get status for pod" podUID="a42aad68-42cf-4264-9c94-8678b0eea1b8" pod="openshift-authentication/oauth-openshift-558db77b4-pmmcw" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/pods/oauth-openshift-558db77b4-pmmcw\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 06 03:38:40 crc kubenswrapper[4980]: I1206 03:38:40.941575 4980 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 06 03:38:40 crc kubenswrapper[4980]: I1206 03:38:40.942354 4980 status_manager.go:851] "Failed to get status for pod" podUID="9c8b3374-77eb-44ad-b270-2263c6c23e28" pod="openshift-marketplace/community-operators-mkd2m" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-mkd2m\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 06 03:38:40 crc kubenswrapper[4980]: I1206 03:38:40.943094 4980 status_manager.go:851] "Failed to get status for pod" podUID="e965bb7a-1bf0-4574-aae4-2a45c75fb11c" pod="openshift-marketplace/certified-operators-58d94" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-58d94\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 06 03:38:40 crc kubenswrapper[4980]: I1206 03:38:40.943738 4980 status_manager.go:851] "Failed to get status for pod" podUID="769abfb5-c18a-4cf0-a033-95b3a808867b" pod="openshift-marketplace/certified-operators-kp68b" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-kp68b\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 06 03:38:40 crc kubenswrapper[4980]: I1206 03:38:40.944346 4980 status_manager.go:851] "Failed to get status for pod" podUID="56a102c7-f341-4076-89dc-44428e77c164" pod="openshift-marketplace/redhat-operators-bl7hx" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-bl7hx\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 06 03:38:40 crc kubenswrapper[4980]: I1206 03:38:40.945026 4980 status_manager.go:851] "Failed to get status for pod" podUID="652a7feb-30e3-4e9f-b1a9-c8ef201c658c" pod="openshift-marketplace/community-operators-2vfzc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-2vfzc\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 06 03:38:40 crc kubenswrapper[4980]: I1206 03:38:40.945550 4980 status_manager.go:851] "Failed to get status for pod" podUID="2ce9bc16-3d2a-4feb-82c1-0f42518196cf" pod="openshift-marketplace/redhat-marketplace-gxcqb" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-gxcqb\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 06 03:38:40 crc kubenswrapper[4980]: I1206 03:38:40.946037 4980 status_manager.go:851] "Failed to get status for pod" podUID="43112506-9fdd-4b19-a5a5-caf864c09774" pod="openshift-marketplace/redhat-operators-m9vl6" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-m9vl6\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 06 03:38:40 crc kubenswrapper[4980]: I1206 03:38:40.946797 4980 status_manager.go:851] "Failed to get status for pod" podUID="3caa51b4-a311-4498-ac9b-10c20989684a" pod="openshift-marketplace/redhat-marketplace-xhnw8" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-xhnw8\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 06 03:38:41 crc kubenswrapper[4980]: I1206 03:38:41.194032 4980 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 06 03:38:41 crc kubenswrapper[4980]: I1206 03:38:41.194712 4980 status_manager.go:851] "Failed to get status for pod" podUID="9c8b3374-77eb-44ad-b270-2263c6c23e28" pod="openshift-marketplace/community-operators-mkd2m" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-mkd2m\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 06 03:38:41 crc kubenswrapper[4980]: I1206 03:38:41.195061 4980 status_manager.go:851] "Failed to get status for pod" podUID="e965bb7a-1bf0-4574-aae4-2a45c75fb11c" pod="openshift-marketplace/certified-operators-58d94" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-58d94\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 06 03:38:41 crc kubenswrapper[4980]: I1206 03:38:41.195670 4980 status_manager.go:851] "Failed to get status for pod" podUID="769abfb5-c18a-4cf0-a033-95b3a808867b" pod="openshift-marketplace/certified-operators-kp68b" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-kp68b\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 06 03:38:41 crc kubenswrapper[4980]: I1206 03:38:41.195985 4980 status_manager.go:851] "Failed to get status for pod" podUID="56a102c7-f341-4076-89dc-44428e77c164" pod="openshift-marketplace/redhat-operators-bl7hx" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-bl7hx\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 06 03:38:41 crc kubenswrapper[4980]: I1206 03:38:41.196234 4980 status_manager.go:851] "Failed to get status for pod" podUID="652a7feb-30e3-4e9f-b1a9-c8ef201c658c" pod="openshift-marketplace/community-operators-2vfzc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-2vfzc\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 06 03:38:41 crc kubenswrapper[4980]: I1206 03:38:41.196589 4980 status_manager.go:851] "Failed to get status for pod" podUID="2ce9bc16-3d2a-4feb-82c1-0f42518196cf" pod="openshift-marketplace/redhat-marketplace-gxcqb" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-gxcqb\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 06 03:38:41 crc kubenswrapper[4980]: I1206 03:38:41.196939 4980 status_manager.go:851] "Failed to get status for pod" podUID="43112506-9fdd-4b19-a5a5-caf864c09774" pod="openshift-marketplace/redhat-operators-m9vl6" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-m9vl6\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 06 03:38:41 crc kubenswrapper[4980]: I1206 03:38:41.197199 4980 status_manager.go:851] "Failed to get status for pod" podUID="3caa51b4-a311-4498-ac9b-10c20989684a" pod="openshift-marketplace/redhat-marketplace-xhnw8" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-xhnw8\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 06 03:38:41 crc kubenswrapper[4980]: I1206 03:38:41.197533 4980 status_manager.go:851] "Failed to get status for pod" podUID="dee63a36-cd1b-410c-b329-bd7966b1d38d" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 06 03:38:41 crc kubenswrapper[4980]: I1206 03:38:41.197840 4980 status_manager.go:851] "Failed to get status for pod" podUID="71bb4a3aecc4ba5b26c4b7318770ce13" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 06 03:38:41 crc kubenswrapper[4980]: I1206 03:38:41.198203 4980 status_manager.go:851] "Failed to get status for pod" podUID="a42aad68-42cf-4264-9c94-8678b0eea1b8" pod="openshift-authentication/oauth-openshift-558db77b4-pmmcw" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/pods/oauth-openshift-558db77b4-pmmcw\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 06 03:38:43 crc kubenswrapper[4980]: I1206 03:38:41.951679 4980 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log" Dec 06 03:38:43 crc kubenswrapper[4980]: I1206 03:38:41.951765 4980 generic.go:334] "Generic (PLEG): container finished" podID="f614b9022728cf315e60c057852e563e" containerID="71f7b492ea07cdf7dc65fc48184ec1e64914eb1811bbfdddc0fb14c8ad6cb296" exitCode=1 Dec 06 03:38:43 crc kubenswrapper[4980]: I1206 03:38:41.951824 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerDied","Data":"71f7b492ea07cdf7dc65fc48184ec1e64914eb1811bbfdddc0fb14c8ad6cb296"} Dec 06 03:38:43 crc kubenswrapper[4980]: I1206 03:38:41.952706 4980 scope.go:117] "RemoveContainer" containerID="71f7b492ea07cdf7dc65fc48184ec1e64914eb1811bbfdddc0fb14c8ad6cb296" Dec 06 03:38:43 crc kubenswrapper[4980]: I1206 03:38:41.953273 4980 status_manager.go:851] "Failed to get status for pod" podUID="3caa51b4-a311-4498-ac9b-10c20989684a" pod="openshift-marketplace/redhat-marketplace-xhnw8" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-xhnw8\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 06 03:38:43 crc kubenswrapper[4980]: I1206 03:38:41.954052 4980 status_manager.go:851] "Failed to get status for pod" podUID="f614b9022728cf315e60c057852e563e" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-controller-manager/pods/kube-controller-manager-crc\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 06 03:38:43 crc kubenswrapper[4980]: I1206 03:38:41.954621 4980 status_manager.go:851] "Failed to get status for pod" podUID="dee63a36-cd1b-410c-b329-bd7966b1d38d" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 06 03:38:43 crc kubenswrapper[4980]: I1206 03:38:41.955030 4980 status_manager.go:851] "Failed to get status for pod" podUID="71bb4a3aecc4ba5b26c4b7318770ce13" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 06 03:38:43 crc kubenswrapper[4980]: I1206 03:38:41.955577 4980 status_manager.go:851] "Failed to get status for pod" podUID="a42aad68-42cf-4264-9c94-8678b0eea1b8" pod="openshift-authentication/oauth-openshift-558db77b4-pmmcw" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/pods/oauth-openshift-558db77b4-pmmcw\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 06 03:38:43 crc kubenswrapper[4980]: I1206 03:38:41.956149 4980 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 06 03:38:43 crc kubenswrapper[4980]: I1206 03:38:41.956648 4980 status_manager.go:851] "Failed to get status for pod" podUID="9c8b3374-77eb-44ad-b270-2263c6c23e28" pod="openshift-marketplace/community-operators-mkd2m" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-mkd2m\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 06 03:38:43 crc kubenswrapper[4980]: I1206 03:38:41.957174 4980 status_manager.go:851] "Failed to get status for pod" podUID="e965bb7a-1bf0-4574-aae4-2a45c75fb11c" pod="openshift-marketplace/certified-operators-58d94" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-58d94\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 06 03:38:43 crc kubenswrapper[4980]: I1206 03:38:41.957684 4980 status_manager.go:851] "Failed to get status for pod" podUID="769abfb5-c18a-4cf0-a033-95b3a808867b" pod="openshift-marketplace/certified-operators-kp68b" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-kp68b\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 06 03:38:43 crc kubenswrapper[4980]: I1206 03:38:41.958116 4980 status_manager.go:851] "Failed to get status for pod" podUID="56a102c7-f341-4076-89dc-44428e77c164" pod="openshift-marketplace/redhat-operators-bl7hx" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-bl7hx\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 06 03:38:43 crc kubenswrapper[4980]: I1206 03:38:41.958587 4980 status_manager.go:851] "Failed to get status for pod" podUID="652a7feb-30e3-4e9f-b1a9-c8ef201c658c" pod="openshift-marketplace/community-operators-2vfzc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-2vfzc\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 06 03:38:43 crc kubenswrapper[4980]: I1206 03:38:41.959149 4980 status_manager.go:851] "Failed to get status for pod" podUID="2ce9bc16-3d2a-4feb-82c1-0f42518196cf" pod="openshift-marketplace/redhat-marketplace-gxcqb" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-gxcqb\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 06 03:38:43 crc kubenswrapper[4980]: I1206 03:38:41.959669 4980 status_manager.go:851] "Failed to get status for pod" podUID="43112506-9fdd-4b19-a5a5-caf864c09774" pod="openshift-marketplace/redhat-operators-m9vl6" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-m9vl6\": dial tcp 38.102.83.184:6443: connect: connection refused" Dec 06 03:38:43 crc kubenswrapper[4980]: I1206 03:38:42.962070 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"1c5fb2e83a299994137324204eaf1daacc8c6ecf6437ba93a5c9e54f791c3d26"} Dec 06 03:38:43 crc kubenswrapper[4980]: I1206 03:38:43.566178 4980 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 06 03:38:43 crc kubenswrapper[4980]: I1206 03:38:43.992590 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"741e95aea0742486336031b2b19155c4128d31db7fbe03f37e92503d8a4a21a6"} Dec 06 03:38:43 crc kubenswrapper[4980]: I1206 03:38:43.995236 4980 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log" Dec 06 03:38:43 crc kubenswrapper[4980]: I1206 03:38:43.995289 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"9b8f83d7e24d41fd3434b725856caa687ef15006ac31546faf238208840d6e15"} Dec 06 03:38:45 crc kubenswrapper[4980]: I1206 03:38:45.003524 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"31bdb246c491fe84446a1ccfca878afaa7bbd2ffc24d248386e1b3cf2821fdd9"} Dec 06 03:38:45 crc kubenswrapper[4980]: I1206 03:38:45.004157 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"f9337da7c835549ed7e69d4703153d2ae8f8f1d2c325339ca717a3494bb2ba0c"} Dec 06 03:38:46 crc kubenswrapper[4980]: I1206 03:38:46.020571 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"0d5f10e779757611e08b970948ebb21b9b0e2a0d0bbb02dba14312e3349d996d"} Dec 06 03:38:46 crc kubenswrapper[4980]: I1206 03:38:46.020872 4980 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 06 03:38:46 crc kubenswrapper[4980]: I1206 03:38:46.021080 4980 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="7f958150-549d-4c80-8a7e-23c39e4f9500" Dec 06 03:38:46 crc kubenswrapper[4980]: I1206 03:38:46.021125 4980 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="7f958150-549d-4c80-8a7e-23c39e4f9500" Dec 06 03:38:46 crc kubenswrapper[4980]: I1206 03:38:46.036805 4980 kubelet.go:1914] "Deleted mirror pod because it is outdated" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 06 03:38:47 crc kubenswrapper[4980]: I1206 03:38:47.027462 4980 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="7f958150-549d-4c80-8a7e-23c39e4f9500" Dec 06 03:38:47 crc kubenswrapper[4980]: I1206 03:38:47.027547 4980 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="7f958150-549d-4c80-8a7e-23c39e4f9500" Dec 06 03:38:48 crc kubenswrapper[4980]: I1206 03:38:48.202780 4980 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 06 03:38:48 crc kubenswrapper[4980]: I1206 03:38:48.203725 4980 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 06 03:38:48 crc kubenswrapper[4980]: I1206 03:38:48.204208 4980 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="7f958150-549d-4c80-8a7e-23c39e4f9500" Dec 06 03:38:48 crc kubenswrapper[4980]: I1206 03:38:48.204232 4980 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="7f958150-549d-4c80-8a7e-23c39e4f9500" Dec 06 03:38:48 crc kubenswrapper[4980]: I1206 03:38:48.208500 4980 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 06 03:38:49 crc kubenswrapper[4980]: I1206 03:38:49.042071 4980 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="7f958150-549d-4c80-8a7e-23c39e4f9500" Dec 06 03:38:49 crc kubenswrapper[4980]: I1206 03:38:49.042131 4980 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="7f958150-549d-4c80-8a7e-23c39e4f9500" Dec 06 03:38:49 crc kubenswrapper[4980]: I1206 03:38:49.049275 4980 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 06 03:38:49 crc kubenswrapper[4980]: I1206 03:38:49.349221 4980 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="71bb4a3aecc4ba5b26c4b7318770ce13" podUID="70cb4ae6-987b-460a-a8a4-31dab396c5d5" Dec 06 03:38:50 crc kubenswrapper[4980]: I1206 03:38:50.047626 4980 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="7f958150-549d-4c80-8a7e-23c39e4f9500" Dec 06 03:38:50 crc kubenswrapper[4980]: I1206 03:38:50.047670 4980 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="7f958150-549d-4c80-8a7e-23c39e4f9500" Dec 06 03:38:50 crc kubenswrapper[4980]: I1206 03:38:50.054434 4980 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="71bb4a3aecc4ba5b26c4b7318770ce13" podUID="70cb4ae6-987b-460a-a8a4-31dab396c5d5" Dec 06 03:38:50 crc kubenswrapper[4980]: I1206 03:38:50.541441 4980 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 06 03:38:50 crc kubenswrapper[4980]: I1206 03:38:50.548822 4980 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 06 03:38:51 crc kubenswrapper[4980]: I1206 03:38:51.052477 4980 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 06 03:38:53 crc kubenswrapper[4980]: I1206 03:38:53.571087 4980 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 06 03:38:58 crc kubenswrapper[4980]: I1206 03:38:58.300580 4980 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Dec 06 03:38:59 crc kubenswrapper[4980]: I1206 03:38:59.349637 4980 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Dec 06 03:38:59 crc kubenswrapper[4980]: I1206 03:38:59.705359 4980 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Dec 06 03:38:59 crc kubenswrapper[4980]: I1206 03:38:59.826475 4980 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Dec 06 03:38:59 crc kubenswrapper[4980]: I1206 03:38:59.946022 4980 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Dec 06 03:39:00 crc kubenswrapper[4980]: I1206 03:39:00.300829 4980 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Dec 06 03:39:00 crc kubenswrapper[4980]: I1206 03:39:00.734416 4980 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Dec 06 03:39:01 crc kubenswrapper[4980]: I1206 03:39:01.114597 4980 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Dec 06 03:39:01 crc kubenswrapper[4980]: I1206 03:39:01.228997 4980 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Dec 06 03:39:01 crc kubenswrapper[4980]: I1206 03:39:01.403058 4980 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Dec 06 03:39:01 crc kubenswrapper[4980]: I1206 03:39:01.414472 4980 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Dec 06 03:39:01 crc kubenswrapper[4980]: I1206 03:39:01.513985 4980 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Dec 06 03:39:01 crc kubenswrapper[4980]: I1206 03:39:01.560900 4980 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Dec 06 03:39:01 crc kubenswrapper[4980]: I1206 03:39:01.563609 4980 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Dec 06 03:39:01 crc kubenswrapper[4980]: I1206 03:39:01.571892 4980 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Dec 06 03:39:01 crc kubenswrapper[4980]: I1206 03:39:01.581096 4980 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Dec 06 03:39:01 crc kubenswrapper[4980]: I1206 03:39:01.748279 4980 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Dec 06 03:39:01 crc kubenswrapper[4980]: I1206 03:39:01.823421 4980 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Dec 06 03:39:02 crc kubenswrapper[4980]: I1206 03:39:02.006760 4980 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Dec 06 03:39:02 crc kubenswrapper[4980]: I1206 03:39:02.120085 4980 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Dec 06 03:39:02 crc kubenswrapper[4980]: I1206 03:39:02.140215 4980 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Dec 06 03:39:02 crc kubenswrapper[4980]: I1206 03:39:02.188247 4980 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Dec 06 03:39:02 crc kubenswrapper[4980]: I1206 03:39:02.230273 4980 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Dec 06 03:39:02 crc kubenswrapper[4980]: I1206 03:39:02.461583 4980 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Dec 06 03:39:02 crc kubenswrapper[4980]: I1206 03:39:02.463935 4980 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Dec 06 03:39:02 crc kubenswrapper[4980]: I1206 03:39:02.473942 4980 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Dec 06 03:39:02 crc kubenswrapper[4980]: I1206 03:39:02.537751 4980 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Dec 06 03:39:02 crc kubenswrapper[4980]: I1206 03:39:02.654684 4980 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Dec 06 03:39:02 crc kubenswrapper[4980]: I1206 03:39:02.735146 4980 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Dec 06 03:39:02 crc kubenswrapper[4980]: I1206 03:39:02.809199 4980 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Dec 06 03:39:02 crc kubenswrapper[4980]: I1206 03:39:02.849943 4980 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Dec 06 03:39:02 crc kubenswrapper[4980]: I1206 03:39:02.910841 4980 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Dec 06 03:39:02 crc kubenswrapper[4980]: I1206 03:39:02.926820 4980 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Dec 06 03:39:02 crc kubenswrapper[4980]: I1206 03:39:02.978219 4980 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Dec 06 03:39:02 crc kubenswrapper[4980]: I1206 03:39:02.987615 4980 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Dec 06 03:39:03 crc kubenswrapper[4980]: I1206 03:39:03.150146 4980 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Dec 06 03:39:03 crc kubenswrapper[4980]: I1206 03:39:03.227462 4980 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Dec 06 03:39:03 crc kubenswrapper[4980]: I1206 03:39:03.255119 4980 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Dec 06 03:39:03 crc kubenswrapper[4980]: I1206 03:39:03.280900 4980 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Dec 06 03:39:03 crc kubenswrapper[4980]: I1206 03:39:03.449352 4980 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Dec 06 03:39:03 crc kubenswrapper[4980]: I1206 03:39:03.488570 4980 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Dec 06 03:39:03 crc kubenswrapper[4980]: I1206 03:39:03.598865 4980 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Dec 06 03:39:03 crc kubenswrapper[4980]: I1206 03:39:03.606773 4980 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Dec 06 03:39:03 crc kubenswrapper[4980]: I1206 03:39:03.622456 4980 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Dec 06 03:39:03 crc kubenswrapper[4980]: I1206 03:39:03.661104 4980 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Dec 06 03:39:03 crc kubenswrapper[4980]: I1206 03:39:03.670149 4980 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Dec 06 03:39:03 crc kubenswrapper[4980]: I1206 03:39:03.704422 4980 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Dec 06 03:39:03 crc kubenswrapper[4980]: I1206 03:39:03.712128 4980 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Dec 06 03:39:03 crc kubenswrapper[4980]: I1206 03:39:03.725844 4980 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Dec 06 03:39:03 crc kubenswrapper[4980]: I1206 03:39:03.741885 4980 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Dec 06 03:39:03 crc kubenswrapper[4980]: I1206 03:39:03.781473 4980 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Dec 06 03:39:03 crc kubenswrapper[4980]: I1206 03:39:03.813723 4980 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Dec 06 03:39:04 crc kubenswrapper[4980]: I1206 03:39:04.061326 4980 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Dec 06 03:39:04 crc kubenswrapper[4980]: I1206 03:39:04.074029 4980 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Dec 06 03:39:04 crc kubenswrapper[4980]: I1206 03:39:04.101382 4980 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Dec 06 03:39:04 crc kubenswrapper[4980]: I1206 03:39:04.117861 4980 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Dec 06 03:39:04 crc kubenswrapper[4980]: I1206 03:39:04.186955 4980 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Dec 06 03:39:04 crc kubenswrapper[4980]: I1206 03:39:04.243583 4980 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Dec 06 03:39:04 crc kubenswrapper[4980]: I1206 03:39:04.288469 4980 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Dec 06 03:39:04 crc kubenswrapper[4980]: I1206 03:39:04.405262 4980 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Dec 06 03:39:04 crc kubenswrapper[4980]: I1206 03:39:04.471952 4980 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Dec 06 03:39:04 crc kubenswrapper[4980]: I1206 03:39:04.473259 4980 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Dec 06 03:39:04 crc kubenswrapper[4980]: I1206 03:39:04.567399 4980 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Dec 06 03:39:04 crc kubenswrapper[4980]: I1206 03:39:04.590692 4980 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Dec 06 03:39:04 crc kubenswrapper[4980]: I1206 03:39:04.654553 4980 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Dec 06 03:39:04 crc kubenswrapper[4980]: I1206 03:39:04.664309 4980 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Dec 06 03:39:04 crc kubenswrapper[4980]: I1206 03:39:04.764052 4980 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Dec 06 03:39:04 crc kubenswrapper[4980]: I1206 03:39:04.994311 4980 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Dec 06 03:39:04 crc kubenswrapper[4980]: I1206 03:39:04.995097 4980 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Dec 06 03:39:05 crc kubenswrapper[4980]: I1206 03:39:05.046066 4980 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Dec 06 03:39:05 crc kubenswrapper[4980]: I1206 03:39:05.284840 4980 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Dec 06 03:39:05 crc kubenswrapper[4980]: I1206 03:39:05.600742 4980 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Dec 06 03:39:05 crc kubenswrapper[4980]: I1206 03:39:05.611272 4980 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Dec 06 03:39:05 crc kubenswrapper[4980]: I1206 03:39:05.647016 4980 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Dec 06 03:39:05 crc kubenswrapper[4980]: I1206 03:39:05.691153 4980 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Dec 06 03:39:05 crc kubenswrapper[4980]: I1206 03:39:05.696874 4980 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Dec 06 03:39:05 crc kubenswrapper[4980]: I1206 03:39:05.715454 4980 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Dec 06 03:39:05 crc kubenswrapper[4980]: I1206 03:39:05.731062 4980 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Dec 06 03:39:05 crc kubenswrapper[4980]: I1206 03:39:05.833543 4980 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Dec 06 03:39:05 crc kubenswrapper[4980]: I1206 03:39:05.837098 4980 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-m9vl6" podStartSLOduration=41.617644908 podStartE2EDuration="2m21.837036289s" podCreationTimestamp="2025-12-06 03:36:44 +0000 UTC" firstStartedPulling="2025-12-06 03:36:47.180861506 +0000 UTC m=+166.425612777" lastFinishedPulling="2025-12-06 03:38:27.400252887 +0000 UTC m=+266.645004158" observedRunningTime="2025-12-06 03:38:49.307407435 +0000 UTC m=+288.552158696" watchObservedRunningTime="2025-12-06 03:39:05.837036289 +0000 UTC m=+305.081787600" Dec 06 03:39:05 crc kubenswrapper[4980]: I1206 03:39:05.839865 4980 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" podStartSLOduration=40.839843791 podStartE2EDuration="40.839843791s" podCreationTimestamp="2025-12-06 03:38:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 03:38:49.140453472 +0000 UTC m=+288.385204773" watchObservedRunningTime="2025-12-06 03:39:05.839843791 +0000 UTC m=+305.084595092" Dec 06 03:39:05 crc kubenswrapper[4980]: I1206 03:39:05.841665 4980 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-2vfzc" podStartSLOduration=41.91058086 podStartE2EDuration="2m23.841651603s" podCreationTimestamp="2025-12-06 03:36:42 +0000 UTC" firstStartedPulling="2025-12-06 03:36:44.920285285 +0000 UTC m=+164.165036556" lastFinishedPulling="2025-12-06 03:38:26.851356018 +0000 UTC m=+266.096107299" observedRunningTime="2025-12-06 03:38:49.275422609 +0000 UTC m=+288.520173890" watchObservedRunningTime="2025-12-06 03:39:05.841651603 +0000 UTC m=+305.086402954" Dec 06 03:39:05 crc kubenswrapper[4980]: I1206 03:39:05.843069 4980 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc","openshift-authentication/oauth-openshift-558db77b4-pmmcw"] Dec 06 03:39:05 crc kubenswrapper[4980]: I1206 03:39:05.843257 4980 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Dec 06 03:39:05 crc kubenswrapper[4980]: I1206 03:39:05.843758 4980 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="7f958150-549d-4c80-8a7e-23c39e4f9500" Dec 06 03:39:05 crc kubenswrapper[4980]: I1206 03:39:05.843811 4980 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="7f958150-549d-4c80-8a7e-23c39e4f9500" Dec 06 03:39:05 crc kubenswrapper[4980]: I1206 03:39:05.845482 4980 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Dec 06 03:39:05 crc kubenswrapper[4980]: I1206 03:39:05.848068 4980 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 06 03:39:05 crc kubenswrapper[4980]: I1206 03:39:05.870618 4980 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=19.870591581 podStartE2EDuration="19.870591581s" podCreationTimestamp="2025-12-06 03:38:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 03:39:05.867289765 +0000 UTC m=+305.112041076" watchObservedRunningTime="2025-12-06 03:39:05.870591581 +0000 UTC m=+305.115342882" Dec 06 03:39:06 crc kubenswrapper[4980]: I1206 03:39:06.124787 4980 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Dec 06 03:39:06 crc kubenswrapper[4980]: I1206 03:39:06.211197 4980 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Dec 06 03:39:06 crc kubenswrapper[4980]: I1206 03:39:06.222432 4980 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Dec 06 03:39:06 crc kubenswrapper[4980]: I1206 03:39:06.401064 4980 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Dec 06 03:39:06 crc kubenswrapper[4980]: I1206 03:39:06.456802 4980 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Dec 06 03:39:06 crc kubenswrapper[4980]: I1206 03:39:06.501593 4980 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Dec 06 03:39:06 crc kubenswrapper[4980]: I1206 03:39:06.620192 4980 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Dec 06 03:39:06 crc kubenswrapper[4980]: I1206 03:39:06.636311 4980 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-6bffc96f45-tkm9h"] Dec 06 03:39:06 crc kubenswrapper[4980]: E1206 03:39:06.636616 4980 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a42aad68-42cf-4264-9c94-8678b0eea1b8" containerName="oauth-openshift" Dec 06 03:39:06 crc kubenswrapper[4980]: I1206 03:39:06.636633 4980 state_mem.go:107] "Deleted CPUSet assignment" podUID="a42aad68-42cf-4264-9c94-8678b0eea1b8" containerName="oauth-openshift" Dec 06 03:39:06 crc kubenswrapper[4980]: E1206 03:39:06.636670 4980 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dee63a36-cd1b-410c-b329-bd7966b1d38d" containerName="installer" Dec 06 03:39:06 crc kubenswrapper[4980]: I1206 03:39:06.636680 4980 state_mem.go:107] "Deleted CPUSet assignment" podUID="dee63a36-cd1b-410c-b329-bd7966b1d38d" containerName="installer" Dec 06 03:39:06 crc kubenswrapper[4980]: I1206 03:39:06.636798 4980 memory_manager.go:354] "RemoveStaleState removing state" podUID="dee63a36-cd1b-410c-b329-bd7966b1d38d" containerName="installer" Dec 06 03:39:06 crc kubenswrapper[4980]: I1206 03:39:06.636811 4980 memory_manager.go:354] "RemoveStaleState removing state" podUID="a42aad68-42cf-4264-9c94-8678b0eea1b8" containerName="oauth-openshift" Dec 06 03:39:06 crc kubenswrapper[4980]: I1206 03:39:06.637278 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-6bffc96f45-tkm9h" Dec 06 03:39:06 crc kubenswrapper[4980]: I1206 03:39:06.639238 4980 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Dec 06 03:39:06 crc kubenswrapper[4980]: I1206 03:39:06.639598 4980 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Dec 06 03:39:06 crc kubenswrapper[4980]: I1206 03:39:06.644758 4980 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Dec 06 03:39:06 crc kubenswrapper[4980]: I1206 03:39:06.644780 4980 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Dec 06 03:39:06 crc kubenswrapper[4980]: I1206 03:39:06.645045 4980 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Dec 06 03:39:06 crc kubenswrapper[4980]: I1206 03:39:06.645067 4980 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Dec 06 03:39:06 crc kubenswrapper[4980]: I1206 03:39:06.644773 4980 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Dec 06 03:39:06 crc kubenswrapper[4980]: I1206 03:39:06.645282 4980 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Dec 06 03:39:06 crc kubenswrapper[4980]: I1206 03:39:06.645288 4980 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Dec 06 03:39:06 crc kubenswrapper[4980]: I1206 03:39:06.645959 4980 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Dec 06 03:39:06 crc kubenswrapper[4980]: I1206 03:39:06.646393 4980 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Dec 06 03:39:06 crc kubenswrapper[4980]: I1206 03:39:06.649297 4980 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Dec 06 03:39:06 crc kubenswrapper[4980]: I1206 03:39:06.651249 4980 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Dec 06 03:39:06 crc kubenswrapper[4980]: I1206 03:39:06.654440 4980 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Dec 06 03:39:06 crc kubenswrapper[4980]: I1206 03:39:06.657700 4980 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Dec 06 03:39:06 crc kubenswrapper[4980]: I1206 03:39:06.675307 4980 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Dec 06 03:39:06 crc kubenswrapper[4980]: I1206 03:39:06.711478 4980 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Dec 06 03:39:06 crc kubenswrapper[4980]: I1206 03:39:06.713978 4980 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Dec 06 03:39:06 crc kubenswrapper[4980]: I1206 03:39:06.762893 4980 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Dec 06 03:39:06 crc kubenswrapper[4980]: I1206 03:39:06.789358 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/3f26f4f4-cf02-4141-964c-4737b34410ed-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-6bffc96f45-tkm9h\" (UID: \"3f26f4f4-cf02-4141-964c-4737b34410ed\") " pod="openshift-authentication/oauth-openshift-6bffc96f45-tkm9h" Dec 06 03:39:06 crc kubenswrapper[4980]: I1206 03:39:06.789396 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xjnd5\" (UniqueName: \"kubernetes.io/projected/3f26f4f4-cf02-4141-964c-4737b34410ed-kube-api-access-xjnd5\") pod \"oauth-openshift-6bffc96f45-tkm9h\" (UID: \"3f26f4f4-cf02-4141-964c-4737b34410ed\") " pod="openshift-authentication/oauth-openshift-6bffc96f45-tkm9h" Dec 06 03:39:06 crc kubenswrapper[4980]: I1206 03:39:06.789428 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/3f26f4f4-cf02-4141-964c-4737b34410ed-v4-0-config-system-serving-cert\") pod \"oauth-openshift-6bffc96f45-tkm9h\" (UID: \"3f26f4f4-cf02-4141-964c-4737b34410ed\") " pod="openshift-authentication/oauth-openshift-6bffc96f45-tkm9h" Dec 06 03:39:06 crc kubenswrapper[4980]: I1206 03:39:06.789462 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/3f26f4f4-cf02-4141-964c-4737b34410ed-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-6bffc96f45-tkm9h\" (UID: \"3f26f4f4-cf02-4141-964c-4737b34410ed\") " pod="openshift-authentication/oauth-openshift-6bffc96f45-tkm9h" Dec 06 03:39:06 crc kubenswrapper[4980]: I1206 03:39:06.789496 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/3f26f4f4-cf02-4141-964c-4737b34410ed-audit-dir\") pod \"oauth-openshift-6bffc96f45-tkm9h\" (UID: \"3f26f4f4-cf02-4141-964c-4737b34410ed\") " pod="openshift-authentication/oauth-openshift-6bffc96f45-tkm9h" Dec 06 03:39:06 crc kubenswrapper[4980]: I1206 03:39:06.789556 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/3f26f4f4-cf02-4141-964c-4737b34410ed-v4-0-config-user-template-error\") pod \"oauth-openshift-6bffc96f45-tkm9h\" (UID: \"3f26f4f4-cf02-4141-964c-4737b34410ed\") " pod="openshift-authentication/oauth-openshift-6bffc96f45-tkm9h" Dec 06 03:39:06 crc kubenswrapper[4980]: I1206 03:39:06.789571 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/3f26f4f4-cf02-4141-964c-4737b34410ed-audit-policies\") pod \"oauth-openshift-6bffc96f45-tkm9h\" (UID: \"3f26f4f4-cf02-4141-964c-4737b34410ed\") " pod="openshift-authentication/oauth-openshift-6bffc96f45-tkm9h" Dec 06 03:39:06 crc kubenswrapper[4980]: I1206 03:39:06.789595 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/3f26f4f4-cf02-4141-964c-4737b34410ed-v4-0-config-system-service-ca\") pod \"oauth-openshift-6bffc96f45-tkm9h\" (UID: \"3f26f4f4-cf02-4141-964c-4737b34410ed\") " pod="openshift-authentication/oauth-openshift-6bffc96f45-tkm9h" Dec 06 03:39:06 crc kubenswrapper[4980]: I1206 03:39:06.789614 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/3f26f4f4-cf02-4141-964c-4737b34410ed-v4-0-config-system-cliconfig\") pod \"oauth-openshift-6bffc96f45-tkm9h\" (UID: \"3f26f4f4-cf02-4141-964c-4737b34410ed\") " pod="openshift-authentication/oauth-openshift-6bffc96f45-tkm9h" Dec 06 03:39:06 crc kubenswrapper[4980]: I1206 03:39:06.789630 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/3f26f4f4-cf02-4141-964c-4737b34410ed-v4-0-config-user-template-login\") pod \"oauth-openshift-6bffc96f45-tkm9h\" (UID: \"3f26f4f4-cf02-4141-964c-4737b34410ed\") " pod="openshift-authentication/oauth-openshift-6bffc96f45-tkm9h" Dec 06 03:39:06 crc kubenswrapper[4980]: I1206 03:39:06.789752 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/3f26f4f4-cf02-4141-964c-4737b34410ed-v4-0-config-system-router-certs\") pod \"oauth-openshift-6bffc96f45-tkm9h\" (UID: \"3f26f4f4-cf02-4141-964c-4737b34410ed\") " pod="openshift-authentication/oauth-openshift-6bffc96f45-tkm9h" Dec 06 03:39:06 crc kubenswrapper[4980]: I1206 03:39:06.789869 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/3f26f4f4-cf02-4141-964c-4737b34410ed-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-6bffc96f45-tkm9h\" (UID: \"3f26f4f4-cf02-4141-964c-4737b34410ed\") " pod="openshift-authentication/oauth-openshift-6bffc96f45-tkm9h" Dec 06 03:39:06 crc kubenswrapper[4980]: I1206 03:39:06.789933 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/3f26f4f4-cf02-4141-964c-4737b34410ed-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-6bffc96f45-tkm9h\" (UID: \"3f26f4f4-cf02-4141-964c-4737b34410ed\") " pod="openshift-authentication/oauth-openshift-6bffc96f45-tkm9h" Dec 06 03:39:06 crc kubenswrapper[4980]: I1206 03:39:06.789986 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/3f26f4f4-cf02-4141-964c-4737b34410ed-v4-0-config-system-session\") pod \"oauth-openshift-6bffc96f45-tkm9h\" (UID: \"3f26f4f4-cf02-4141-964c-4737b34410ed\") " pod="openshift-authentication/oauth-openshift-6bffc96f45-tkm9h" Dec 06 03:39:06 crc kubenswrapper[4980]: I1206 03:39:06.876606 4980 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Dec 06 03:39:06 crc kubenswrapper[4980]: I1206 03:39:06.891067 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/3f26f4f4-cf02-4141-964c-4737b34410ed-v4-0-config-system-cliconfig\") pod \"oauth-openshift-6bffc96f45-tkm9h\" (UID: \"3f26f4f4-cf02-4141-964c-4737b34410ed\") " pod="openshift-authentication/oauth-openshift-6bffc96f45-tkm9h" Dec 06 03:39:06 crc kubenswrapper[4980]: I1206 03:39:06.891160 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/3f26f4f4-cf02-4141-964c-4737b34410ed-v4-0-config-user-template-login\") pod \"oauth-openshift-6bffc96f45-tkm9h\" (UID: \"3f26f4f4-cf02-4141-964c-4737b34410ed\") " pod="openshift-authentication/oauth-openshift-6bffc96f45-tkm9h" Dec 06 03:39:06 crc kubenswrapper[4980]: I1206 03:39:06.891219 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/3f26f4f4-cf02-4141-964c-4737b34410ed-v4-0-config-system-router-certs\") pod \"oauth-openshift-6bffc96f45-tkm9h\" (UID: \"3f26f4f4-cf02-4141-964c-4737b34410ed\") " pod="openshift-authentication/oauth-openshift-6bffc96f45-tkm9h" Dec 06 03:39:06 crc kubenswrapper[4980]: I1206 03:39:06.891282 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/3f26f4f4-cf02-4141-964c-4737b34410ed-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-6bffc96f45-tkm9h\" (UID: \"3f26f4f4-cf02-4141-964c-4737b34410ed\") " pod="openshift-authentication/oauth-openshift-6bffc96f45-tkm9h" Dec 06 03:39:06 crc kubenswrapper[4980]: I1206 03:39:06.891345 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/3f26f4f4-cf02-4141-964c-4737b34410ed-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-6bffc96f45-tkm9h\" (UID: \"3f26f4f4-cf02-4141-964c-4737b34410ed\") " pod="openshift-authentication/oauth-openshift-6bffc96f45-tkm9h" Dec 06 03:39:06 crc kubenswrapper[4980]: I1206 03:39:06.891414 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/3f26f4f4-cf02-4141-964c-4737b34410ed-v4-0-config-system-session\") pod \"oauth-openshift-6bffc96f45-tkm9h\" (UID: \"3f26f4f4-cf02-4141-964c-4737b34410ed\") " pod="openshift-authentication/oauth-openshift-6bffc96f45-tkm9h" Dec 06 03:39:06 crc kubenswrapper[4980]: I1206 03:39:06.891551 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xjnd5\" (UniqueName: \"kubernetes.io/projected/3f26f4f4-cf02-4141-964c-4737b34410ed-kube-api-access-xjnd5\") pod \"oauth-openshift-6bffc96f45-tkm9h\" (UID: \"3f26f4f4-cf02-4141-964c-4737b34410ed\") " pod="openshift-authentication/oauth-openshift-6bffc96f45-tkm9h" Dec 06 03:39:06 crc kubenswrapper[4980]: I1206 03:39:06.891605 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/3f26f4f4-cf02-4141-964c-4737b34410ed-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-6bffc96f45-tkm9h\" (UID: \"3f26f4f4-cf02-4141-964c-4737b34410ed\") " pod="openshift-authentication/oauth-openshift-6bffc96f45-tkm9h" Dec 06 03:39:06 crc kubenswrapper[4980]: I1206 03:39:06.891680 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/3f26f4f4-cf02-4141-964c-4737b34410ed-v4-0-config-system-serving-cert\") pod \"oauth-openshift-6bffc96f45-tkm9h\" (UID: \"3f26f4f4-cf02-4141-964c-4737b34410ed\") " pod="openshift-authentication/oauth-openshift-6bffc96f45-tkm9h" Dec 06 03:39:06 crc kubenswrapper[4980]: I1206 03:39:06.891747 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/3f26f4f4-cf02-4141-964c-4737b34410ed-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-6bffc96f45-tkm9h\" (UID: \"3f26f4f4-cf02-4141-964c-4737b34410ed\") " pod="openshift-authentication/oauth-openshift-6bffc96f45-tkm9h" Dec 06 03:39:06 crc kubenswrapper[4980]: I1206 03:39:06.891843 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/3f26f4f4-cf02-4141-964c-4737b34410ed-audit-dir\") pod \"oauth-openshift-6bffc96f45-tkm9h\" (UID: \"3f26f4f4-cf02-4141-964c-4737b34410ed\") " pod="openshift-authentication/oauth-openshift-6bffc96f45-tkm9h" Dec 06 03:39:06 crc kubenswrapper[4980]: I1206 03:39:06.891898 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/3f26f4f4-cf02-4141-964c-4737b34410ed-v4-0-config-user-template-error\") pod \"oauth-openshift-6bffc96f45-tkm9h\" (UID: \"3f26f4f4-cf02-4141-964c-4737b34410ed\") " pod="openshift-authentication/oauth-openshift-6bffc96f45-tkm9h" Dec 06 03:39:06 crc kubenswrapper[4980]: I1206 03:39:06.891947 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/3f26f4f4-cf02-4141-964c-4737b34410ed-audit-policies\") pod \"oauth-openshift-6bffc96f45-tkm9h\" (UID: \"3f26f4f4-cf02-4141-964c-4737b34410ed\") " pod="openshift-authentication/oauth-openshift-6bffc96f45-tkm9h" Dec 06 03:39:06 crc kubenswrapper[4980]: I1206 03:39:06.892023 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/3f26f4f4-cf02-4141-964c-4737b34410ed-v4-0-config-system-service-ca\") pod \"oauth-openshift-6bffc96f45-tkm9h\" (UID: \"3f26f4f4-cf02-4141-964c-4737b34410ed\") " pod="openshift-authentication/oauth-openshift-6bffc96f45-tkm9h" Dec 06 03:39:06 crc kubenswrapper[4980]: I1206 03:39:06.892144 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/3f26f4f4-cf02-4141-964c-4737b34410ed-v4-0-config-system-cliconfig\") pod \"oauth-openshift-6bffc96f45-tkm9h\" (UID: \"3f26f4f4-cf02-4141-964c-4737b34410ed\") " pod="openshift-authentication/oauth-openshift-6bffc96f45-tkm9h" Dec 06 03:39:06 crc kubenswrapper[4980]: I1206 03:39:06.893040 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/3f26f4f4-cf02-4141-964c-4737b34410ed-audit-dir\") pod \"oauth-openshift-6bffc96f45-tkm9h\" (UID: \"3f26f4f4-cf02-4141-964c-4737b34410ed\") " pod="openshift-authentication/oauth-openshift-6bffc96f45-tkm9h" Dec 06 03:39:06 crc kubenswrapper[4980]: I1206 03:39:06.893263 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/3f26f4f4-cf02-4141-964c-4737b34410ed-v4-0-config-system-service-ca\") pod \"oauth-openshift-6bffc96f45-tkm9h\" (UID: \"3f26f4f4-cf02-4141-964c-4737b34410ed\") " pod="openshift-authentication/oauth-openshift-6bffc96f45-tkm9h" Dec 06 03:39:06 crc kubenswrapper[4980]: I1206 03:39:06.894091 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/3f26f4f4-cf02-4141-964c-4737b34410ed-audit-policies\") pod \"oauth-openshift-6bffc96f45-tkm9h\" (UID: \"3f26f4f4-cf02-4141-964c-4737b34410ed\") " pod="openshift-authentication/oauth-openshift-6bffc96f45-tkm9h" Dec 06 03:39:06 crc kubenswrapper[4980]: I1206 03:39:06.894234 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/3f26f4f4-cf02-4141-964c-4737b34410ed-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-6bffc96f45-tkm9h\" (UID: \"3f26f4f4-cf02-4141-964c-4737b34410ed\") " pod="openshift-authentication/oauth-openshift-6bffc96f45-tkm9h" Dec 06 03:39:06 crc kubenswrapper[4980]: I1206 03:39:06.899977 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/3f26f4f4-cf02-4141-964c-4737b34410ed-v4-0-config-system-session\") pod \"oauth-openshift-6bffc96f45-tkm9h\" (UID: \"3f26f4f4-cf02-4141-964c-4737b34410ed\") " pod="openshift-authentication/oauth-openshift-6bffc96f45-tkm9h" Dec 06 03:39:06 crc kubenswrapper[4980]: I1206 03:39:06.900015 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/3f26f4f4-cf02-4141-964c-4737b34410ed-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-6bffc96f45-tkm9h\" (UID: \"3f26f4f4-cf02-4141-964c-4737b34410ed\") " pod="openshift-authentication/oauth-openshift-6bffc96f45-tkm9h" Dec 06 03:39:06 crc kubenswrapper[4980]: I1206 03:39:06.901444 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/3f26f4f4-cf02-4141-964c-4737b34410ed-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-6bffc96f45-tkm9h\" (UID: \"3f26f4f4-cf02-4141-964c-4737b34410ed\") " pod="openshift-authentication/oauth-openshift-6bffc96f45-tkm9h" Dec 06 03:39:06 crc kubenswrapper[4980]: I1206 03:39:06.903101 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/3f26f4f4-cf02-4141-964c-4737b34410ed-v4-0-config-system-serving-cert\") pod \"oauth-openshift-6bffc96f45-tkm9h\" (UID: \"3f26f4f4-cf02-4141-964c-4737b34410ed\") " pod="openshift-authentication/oauth-openshift-6bffc96f45-tkm9h" Dec 06 03:39:06 crc kubenswrapper[4980]: I1206 03:39:06.904983 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/3f26f4f4-cf02-4141-964c-4737b34410ed-v4-0-config-user-template-error\") pod \"oauth-openshift-6bffc96f45-tkm9h\" (UID: \"3f26f4f4-cf02-4141-964c-4737b34410ed\") " pod="openshift-authentication/oauth-openshift-6bffc96f45-tkm9h" Dec 06 03:39:06 crc kubenswrapper[4980]: I1206 03:39:06.906354 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/3f26f4f4-cf02-4141-964c-4737b34410ed-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-6bffc96f45-tkm9h\" (UID: \"3f26f4f4-cf02-4141-964c-4737b34410ed\") " pod="openshift-authentication/oauth-openshift-6bffc96f45-tkm9h" Dec 06 03:39:06 crc kubenswrapper[4980]: I1206 03:39:06.908173 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/3f26f4f4-cf02-4141-964c-4737b34410ed-v4-0-config-system-router-certs\") pod \"oauth-openshift-6bffc96f45-tkm9h\" (UID: \"3f26f4f4-cf02-4141-964c-4737b34410ed\") " pod="openshift-authentication/oauth-openshift-6bffc96f45-tkm9h" Dec 06 03:39:06 crc kubenswrapper[4980]: I1206 03:39:06.908630 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/3f26f4f4-cf02-4141-964c-4737b34410ed-v4-0-config-user-template-login\") pod \"oauth-openshift-6bffc96f45-tkm9h\" (UID: \"3f26f4f4-cf02-4141-964c-4737b34410ed\") " pod="openshift-authentication/oauth-openshift-6bffc96f45-tkm9h" Dec 06 03:39:06 crc kubenswrapper[4980]: I1206 03:39:06.924883 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xjnd5\" (UniqueName: \"kubernetes.io/projected/3f26f4f4-cf02-4141-964c-4737b34410ed-kube-api-access-xjnd5\") pod \"oauth-openshift-6bffc96f45-tkm9h\" (UID: \"3f26f4f4-cf02-4141-964c-4737b34410ed\") " pod="openshift-authentication/oauth-openshift-6bffc96f45-tkm9h" Dec 06 03:39:06 crc kubenswrapper[4980]: I1206 03:39:06.948201 4980 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Dec 06 03:39:06 crc kubenswrapper[4980]: I1206 03:39:06.961351 4980 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Dec 06 03:39:06 crc kubenswrapper[4980]: I1206 03:39:06.971823 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-6bffc96f45-tkm9h" Dec 06 03:39:06 crc kubenswrapper[4980]: I1206 03:39:06.983627 4980 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Dec 06 03:39:07 crc kubenswrapper[4980]: I1206 03:39:07.046119 4980 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Dec 06 03:39:07 crc kubenswrapper[4980]: I1206 03:39:07.053904 4980 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Dec 06 03:39:07 crc kubenswrapper[4980]: I1206 03:39:07.149251 4980 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Dec 06 03:39:07 crc kubenswrapper[4980]: I1206 03:39:07.191904 4980 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a42aad68-42cf-4264-9c94-8678b0eea1b8" path="/var/lib/kubelet/pods/a42aad68-42cf-4264-9c94-8678b0eea1b8/volumes" Dec 06 03:39:07 crc kubenswrapper[4980]: I1206 03:39:07.277300 4980 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Dec 06 03:39:07 crc kubenswrapper[4980]: I1206 03:39:07.298746 4980 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Dec 06 03:39:07 crc kubenswrapper[4980]: I1206 03:39:07.336714 4980 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Dec 06 03:39:07 crc kubenswrapper[4980]: I1206 03:39:07.468728 4980 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Dec 06 03:39:07 crc kubenswrapper[4980]: I1206 03:39:07.522193 4980 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Dec 06 03:39:07 crc kubenswrapper[4980]: I1206 03:39:07.632253 4980 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Dec 06 03:39:07 crc kubenswrapper[4980]: I1206 03:39:07.680028 4980 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Dec 06 03:39:07 crc kubenswrapper[4980]: I1206 03:39:07.712999 4980 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Dec 06 03:39:07 crc kubenswrapper[4980]: I1206 03:39:07.740031 4980 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Dec 06 03:39:07 crc kubenswrapper[4980]: I1206 03:39:07.757695 4980 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Dec 06 03:39:07 crc kubenswrapper[4980]: I1206 03:39:07.819365 4980 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Dec 06 03:39:08 crc kubenswrapper[4980]: I1206 03:39:08.011046 4980 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Dec 06 03:39:08 crc kubenswrapper[4980]: I1206 03:39:08.026962 4980 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Dec 06 03:39:08 crc kubenswrapper[4980]: I1206 03:39:08.053715 4980 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Dec 06 03:39:08 crc kubenswrapper[4980]: I1206 03:39:08.056974 4980 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Dec 06 03:39:08 crc kubenswrapper[4980]: I1206 03:39:08.106341 4980 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Dec 06 03:39:08 crc kubenswrapper[4980]: I1206 03:39:08.196465 4980 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Dec 06 03:39:08 crc kubenswrapper[4980]: I1206 03:39:08.237801 4980 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Dec 06 03:39:08 crc kubenswrapper[4980]: I1206 03:39:08.258962 4980 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Dec 06 03:39:08 crc kubenswrapper[4980]: I1206 03:39:08.334381 4980 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Dec 06 03:39:08 crc kubenswrapper[4980]: I1206 03:39:08.379580 4980 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Dec 06 03:39:08 crc kubenswrapper[4980]: I1206 03:39:08.407104 4980 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Dec 06 03:39:08 crc kubenswrapper[4980]: I1206 03:39:08.476782 4980 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Dec 06 03:39:08 crc kubenswrapper[4980]: I1206 03:39:08.545995 4980 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Dec 06 03:39:08 crc kubenswrapper[4980]: I1206 03:39:08.624751 4980 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Dec 06 03:39:08 crc kubenswrapper[4980]: I1206 03:39:08.650956 4980 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Dec 06 03:39:08 crc kubenswrapper[4980]: I1206 03:39:08.692590 4980 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Dec 06 03:39:08 crc kubenswrapper[4980]: I1206 03:39:08.707801 4980 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Dec 06 03:39:08 crc kubenswrapper[4980]: I1206 03:39:08.830674 4980 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Dec 06 03:39:08 crc kubenswrapper[4980]: I1206 03:39:08.978022 4980 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Dec 06 03:39:09 crc kubenswrapper[4980]: I1206 03:39:09.287857 4980 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Dec 06 03:39:09 crc kubenswrapper[4980]: I1206 03:39:09.311179 4980 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Dec 06 03:39:09 crc kubenswrapper[4980]: I1206 03:39:09.342090 4980 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Dec 06 03:39:09 crc kubenswrapper[4980]: I1206 03:39:09.348399 4980 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Dec 06 03:39:09 crc kubenswrapper[4980]: I1206 03:39:09.445705 4980 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Dec 06 03:39:09 crc kubenswrapper[4980]: I1206 03:39:09.522438 4980 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Dec 06 03:39:09 crc kubenswrapper[4980]: I1206 03:39:09.533319 4980 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Dec 06 03:39:09 crc kubenswrapper[4980]: I1206 03:39:09.652382 4980 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Dec 06 03:39:09 crc kubenswrapper[4980]: I1206 03:39:09.698544 4980 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Dec 06 03:39:09 crc kubenswrapper[4980]: I1206 03:39:09.774161 4980 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Dec 06 03:39:09 crc kubenswrapper[4980]: I1206 03:39:09.777304 4980 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Dec 06 03:39:09 crc kubenswrapper[4980]: I1206 03:39:09.786528 4980 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Dec 06 03:39:09 crc kubenswrapper[4980]: I1206 03:39:09.985153 4980 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Dec 06 03:39:10 crc kubenswrapper[4980]: I1206 03:39:10.002526 4980 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Dec 06 03:39:10 crc kubenswrapper[4980]: I1206 03:39:10.111714 4980 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Dec 06 03:39:10 crc kubenswrapper[4980]: I1206 03:39:10.202119 4980 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Dec 06 03:39:10 crc kubenswrapper[4980]: I1206 03:39:10.229069 4980 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Dec 06 03:39:10 crc kubenswrapper[4980]: I1206 03:39:10.242441 4980 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Dec 06 03:39:10 crc kubenswrapper[4980]: I1206 03:39:10.256533 4980 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Dec 06 03:39:10 crc kubenswrapper[4980]: I1206 03:39:10.282648 4980 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Dec 06 03:39:10 crc kubenswrapper[4980]: I1206 03:39:10.340131 4980 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Dec 06 03:39:10 crc kubenswrapper[4980]: I1206 03:39:10.343986 4980 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Dec 06 03:39:10 crc kubenswrapper[4980]: I1206 03:39:10.349929 4980 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Dec 06 03:39:10 crc kubenswrapper[4980]: I1206 03:39:10.385493 4980 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Dec 06 03:39:10 crc kubenswrapper[4980]: I1206 03:39:10.421014 4980 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Dec 06 03:39:10 crc kubenswrapper[4980]: I1206 03:39:10.456166 4980 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Dec 06 03:39:10 crc kubenswrapper[4980]: I1206 03:39:10.524672 4980 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Dec 06 03:39:10 crc kubenswrapper[4980]: I1206 03:39:10.563944 4980 kubelet.go:2431] "SyncLoop REMOVE" source="file" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Dec 06 03:39:10 crc kubenswrapper[4980]: I1206 03:39:10.564182 4980 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" containerID="cri-o://86f06d4f20478f6fc3524bcb449288b21a827bc591d313aecbb0772ed610d129" gracePeriod=5 Dec 06 03:39:10 crc kubenswrapper[4980]: I1206 03:39:10.576032 4980 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Dec 06 03:39:10 crc kubenswrapper[4980]: I1206 03:39:10.585225 4980 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Dec 06 03:39:10 crc kubenswrapper[4980]: I1206 03:39:10.790995 4980 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Dec 06 03:39:10 crc kubenswrapper[4980]: I1206 03:39:10.824667 4980 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Dec 06 03:39:10 crc kubenswrapper[4980]: I1206 03:39:10.846639 4980 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Dec 06 03:39:10 crc kubenswrapper[4980]: I1206 03:39:10.971143 4980 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Dec 06 03:39:11 crc kubenswrapper[4980]: I1206 03:39:11.090373 4980 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Dec 06 03:39:11 crc kubenswrapper[4980]: I1206 03:39:11.115187 4980 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Dec 06 03:39:11 crc kubenswrapper[4980]: I1206 03:39:11.177554 4980 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Dec 06 03:39:11 crc kubenswrapper[4980]: I1206 03:39:11.187566 4980 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Dec 06 03:39:11 crc kubenswrapper[4980]: I1206 03:39:11.293259 4980 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Dec 06 03:39:11 crc kubenswrapper[4980]: I1206 03:39:11.339095 4980 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Dec 06 03:39:11 crc kubenswrapper[4980]: I1206 03:39:11.380614 4980 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Dec 06 03:39:11 crc kubenswrapper[4980]: I1206 03:39:11.394091 4980 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Dec 06 03:39:11 crc kubenswrapper[4980]: I1206 03:39:11.413714 4980 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Dec 06 03:39:11 crc kubenswrapper[4980]: I1206 03:39:11.423828 4980 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Dec 06 03:39:11 crc kubenswrapper[4980]: I1206 03:39:11.520286 4980 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Dec 06 03:39:11 crc kubenswrapper[4980]: I1206 03:39:11.534184 4980 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Dec 06 03:39:11 crc kubenswrapper[4980]: I1206 03:39:11.578319 4980 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Dec 06 03:39:11 crc kubenswrapper[4980]: I1206 03:39:11.660153 4980 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Dec 06 03:39:11 crc kubenswrapper[4980]: I1206 03:39:11.713824 4980 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Dec 06 03:39:11 crc kubenswrapper[4980]: I1206 03:39:11.724012 4980 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Dec 06 03:39:11 crc kubenswrapper[4980]: I1206 03:39:11.762127 4980 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Dec 06 03:39:11 crc kubenswrapper[4980]: I1206 03:39:11.781019 4980 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Dec 06 03:39:11 crc kubenswrapper[4980]: I1206 03:39:11.822634 4980 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Dec 06 03:39:11 crc kubenswrapper[4980]: I1206 03:39:11.823469 4980 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Dec 06 03:39:11 crc kubenswrapper[4980]: I1206 03:39:11.824186 4980 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Dec 06 03:39:11 crc kubenswrapper[4980]: I1206 03:39:11.828044 4980 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Dec 06 03:39:11 crc kubenswrapper[4980]: I1206 03:39:11.964112 4980 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Dec 06 03:39:11 crc kubenswrapper[4980]: I1206 03:39:11.973812 4980 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Dec 06 03:39:11 crc kubenswrapper[4980]: I1206 03:39:11.975788 4980 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Dec 06 03:39:12 crc kubenswrapper[4980]: I1206 03:39:12.018572 4980 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Dec 06 03:39:12 crc kubenswrapper[4980]: I1206 03:39:12.022676 4980 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Dec 06 03:39:12 crc kubenswrapper[4980]: I1206 03:39:12.046374 4980 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Dec 06 03:39:12 crc kubenswrapper[4980]: I1206 03:39:12.253892 4980 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Dec 06 03:39:12 crc kubenswrapper[4980]: I1206 03:39:12.289942 4980 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Dec 06 03:39:12 crc kubenswrapper[4980]: I1206 03:39:12.302784 4980 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Dec 06 03:39:12 crc kubenswrapper[4980]: I1206 03:39:12.381943 4980 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Dec 06 03:39:12 crc kubenswrapper[4980]: I1206 03:39:12.463619 4980 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Dec 06 03:39:12 crc kubenswrapper[4980]: I1206 03:39:12.509322 4980 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Dec 06 03:39:12 crc kubenswrapper[4980]: I1206 03:39:12.609599 4980 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Dec 06 03:39:12 crc kubenswrapper[4980]: I1206 03:39:12.647948 4980 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Dec 06 03:39:12 crc kubenswrapper[4980]: I1206 03:39:12.749502 4980 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Dec 06 03:39:12 crc kubenswrapper[4980]: I1206 03:39:12.769068 4980 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Dec 06 03:39:12 crc kubenswrapper[4980]: I1206 03:39:12.834275 4980 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Dec 06 03:39:13 crc kubenswrapper[4980]: I1206 03:39:13.310012 4980 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Dec 06 03:39:13 crc kubenswrapper[4980]: I1206 03:39:13.471867 4980 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Dec 06 03:39:13 crc kubenswrapper[4980]: I1206 03:39:13.474772 4980 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Dec 06 03:39:13 crc kubenswrapper[4980]: I1206 03:39:13.476975 4980 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Dec 06 03:39:13 crc kubenswrapper[4980]: I1206 03:39:13.532182 4980 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Dec 06 03:39:13 crc kubenswrapper[4980]: I1206 03:39:13.593464 4980 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Dec 06 03:39:13 crc kubenswrapper[4980]: I1206 03:39:13.710812 4980 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Dec 06 03:39:13 crc kubenswrapper[4980]: I1206 03:39:13.749782 4980 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Dec 06 03:39:13 crc kubenswrapper[4980]: I1206 03:39:13.774074 4980 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Dec 06 03:39:13 crc kubenswrapper[4980]: I1206 03:39:13.844166 4980 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Dec 06 03:39:14 crc kubenswrapper[4980]: I1206 03:39:14.049856 4980 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Dec 06 03:39:14 crc kubenswrapper[4980]: I1206 03:39:14.057283 4980 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Dec 06 03:39:14 crc kubenswrapper[4980]: I1206 03:39:14.280312 4980 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Dec 06 03:39:14 crc kubenswrapper[4980]: I1206 03:39:14.388901 4980 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Dec 06 03:39:14 crc kubenswrapper[4980]: I1206 03:39:14.479400 4980 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Dec 06 03:39:14 crc kubenswrapper[4980]: I1206 03:39:14.666130 4980 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Dec 06 03:39:15 crc kubenswrapper[4980]: I1206 03:39:15.341994 4980 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Dec 06 03:39:16 crc kubenswrapper[4980]: I1206 03:39:16.154238 4980 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Dec 06 03:39:16 crc kubenswrapper[4980]: I1206 03:39:16.154629 4980 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 06 03:39:16 crc kubenswrapper[4980]: I1206 03:39:16.224370 4980 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Dec 06 03:39:16 crc kubenswrapper[4980]: I1206 03:39:16.224477 4980 generic.go:334] "Generic (PLEG): container finished" podID="f85e55b1a89d02b0cb034b1ea31ed45a" containerID="86f06d4f20478f6fc3524bcb449288b21a827bc591d313aecbb0772ed610d129" exitCode=137 Dec 06 03:39:16 crc kubenswrapper[4980]: I1206 03:39:16.224666 4980 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 06 03:39:16 crc kubenswrapper[4980]: I1206 03:39:16.224699 4980 scope.go:117] "RemoveContainer" containerID="86f06d4f20478f6fc3524bcb449288b21a827bc591d313aecbb0772ed610d129" Dec 06 03:39:16 crc kubenswrapper[4980]: I1206 03:39:16.250124 4980 scope.go:117] "RemoveContainer" containerID="86f06d4f20478f6fc3524bcb449288b21a827bc591d313aecbb0772ed610d129" Dec 06 03:39:16 crc kubenswrapper[4980]: E1206 03:39:16.251035 4980 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"86f06d4f20478f6fc3524bcb449288b21a827bc591d313aecbb0772ed610d129\": container with ID starting with 86f06d4f20478f6fc3524bcb449288b21a827bc591d313aecbb0772ed610d129 not found: ID does not exist" containerID="86f06d4f20478f6fc3524bcb449288b21a827bc591d313aecbb0772ed610d129" Dec 06 03:39:16 crc kubenswrapper[4980]: I1206 03:39:16.251094 4980 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"86f06d4f20478f6fc3524bcb449288b21a827bc591d313aecbb0772ed610d129"} err="failed to get container status \"86f06d4f20478f6fc3524bcb449288b21a827bc591d313aecbb0772ed610d129\": rpc error: code = NotFound desc = could not find container \"86f06d4f20478f6fc3524bcb449288b21a827bc591d313aecbb0772ed610d129\": container with ID starting with 86f06d4f20478f6fc3524bcb449288b21a827bc591d313aecbb0772ed610d129 not found: ID does not exist" Dec 06 03:39:16 crc kubenswrapper[4980]: I1206 03:39:16.334201 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Dec 06 03:39:16 crc kubenswrapper[4980]: I1206 03:39:16.334382 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Dec 06 03:39:16 crc kubenswrapper[4980]: I1206 03:39:16.334431 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Dec 06 03:39:16 crc kubenswrapper[4980]: I1206 03:39:16.334503 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Dec 06 03:39:16 crc kubenswrapper[4980]: I1206 03:39:16.334612 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Dec 06 03:39:16 crc kubenswrapper[4980]: I1206 03:39:16.334630 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock" (OuterVolumeSpecName: "var-lock") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "var-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 03:39:16 crc kubenswrapper[4980]: I1206 03:39:16.334756 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log" (OuterVolumeSpecName: "var-log") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "var-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 03:39:16 crc kubenswrapper[4980]: I1206 03:39:16.334759 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests" (OuterVolumeSpecName: "manifests") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "manifests". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 03:39:16 crc kubenswrapper[4980]: I1206 03:39:16.335236 4980 reconciler_common.go:293] "Volume detached for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") on node \"crc\" DevicePath \"\"" Dec 06 03:39:16 crc kubenswrapper[4980]: I1206 03:39:16.335276 4980 reconciler_common.go:293] "Volume detached for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") on node \"crc\" DevicePath \"\"" Dec 06 03:39:16 crc kubenswrapper[4980]: I1206 03:39:16.335295 4980 reconciler_common.go:293] "Volume detached for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") on node \"crc\" DevicePath \"\"" Dec 06 03:39:16 crc kubenswrapper[4980]: I1206 03:39:16.335327 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir" (OuterVolumeSpecName: "resource-dir") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 03:39:16 crc kubenswrapper[4980]: I1206 03:39:16.351844 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir" (OuterVolumeSpecName: "pod-resource-dir") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "pod-resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 03:39:16 crc kubenswrapper[4980]: I1206 03:39:16.436430 4980 reconciler_common.go:293] "Volume detached for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") on node \"crc\" DevicePath \"\"" Dec 06 03:39:16 crc kubenswrapper[4980]: I1206 03:39:16.436471 4980 reconciler_common.go:293] "Volume detached for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") on node \"crc\" DevicePath \"\"" Dec 06 03:39:17 crc kubenswrapper[4980]: I1206 03:39:17.209467 4980 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" path="/var/lib/kubelet/pods/f85e55b1a89d02b0cb034b1ea31ed45a/volumes" Dec 06 03:39:17 crc kubenswrapper[4980]: I1206 03:39:17.210159 4980 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" podUID="" Dec 06 03:39:17 crc kubenswrapper[4980]: I1206 03:39:17.227259 4980 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Dec 06 03:39:17 crc kubenswrapper[4980]: I1206 03:39:17.227307 4980 kubelet.go:2649] "Unable to find pod for mirror pod, skipping" mirrorPod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" mirrorPodUID="f5e3264a-4820-4dd5-b584-de801751876b" Dec 06 03:39:17 crc kubenswrapper[4980]: I1206 03:39:17.248680 4980 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Dec 06 03:39:17 crc kubenswrapper[4980]: I1206 03:39:17.248745 4980 kubelet.go:2673] "Unable to find pod for mirror pod, skipping" mirrorPod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" mirrorPodUID="f5e3264a-4820-4dd5-b584-de801751876b" Dec 06 03:39:25 crc kubenswrapper[4980]: I1206 03:39:25.338190 4980 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Dec 06 03:39:25 crc kubenswrapper[4980]: I1206 03:39:25.834403 4980 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Dec 06 03:39:26 crc kubenswrapper[4980]: I1206 03:39:26.168938 4980 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Dec 06 03:39:31 crc kubenswrapper[4980]: I1206 03:39:31.720756 4980 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Dec 06 03:39:31 crc kubenswrapper[4980]: I1206 03:39:31.825940 4980 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Dec 06 03:39:32 crc kubenswrapper[4980]: I1206 03:39:32.823328 4980 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Dec 06 03:39:34 crc kubenswrapper[4980]: I1206 03:39:34.058682 4980 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Dec 06 03:39:34 crc kubenswrapper[4980]: I1206 03:39:34.474411 4980 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Dec 06 03:39:39 crc kubenswrapper[4980]: I1206 03:39:39.085796 4980 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Dec 06 03:39:39 crc kubenswrapper[4980]: I1206 03:39:39.253892 4980 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Dec 06 03:39:40 crc kubenswrapper[4980]: I1206 03:39:40.394388 4980 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Dec 06 03:39:40 crc kubenswrapper[4980]: I1206 03:39:40.425151 4980 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Dec 06 03:39:41 crc kubenswrapper[4980]: I1206 03:39:41.287031 4980 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Dec 06 03:39:41 crc kubenswrapper[4980]: I1206 03:39:41.395338 4980 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-6bffc96f45-tkm9h"] Dec 06 03:39:41 crc kubenswrapper[4980]: I1206 03:39:41.625902 4980 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Dec 06 03:39:41 crc kubenswrapper[4980]: I1206 03:39:41.644587 4980 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-6bffc96f45-tkm9h"] Dec 06 03:39:41 crc kubenswrapper[4980]: I1206 03:39:41.712463 4980 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Dec 06 03:39:42 crc kubenswrapper[4980]: I1206 03:39:42.363894 4980 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Dec 06 03:39:42 crc kubenswrapper[4980]: I1206 03:39:42.421539 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-6bffc96f45-tkm9h" event={"ID":"3f26f4f4-cf02-4141-964c-4737b34410ed","Type":"ContainerStarted","Data":"cc1eccf08c49e40f378898781aa207de77b5b300e47216e5faece8d68fb9ac8c"} Dec 06 03:39:42 crc kubenswrapper[4980]: I1206 03:39:42.421581 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-6bffc96f45-tkm9h" event={"ID":"3f26f4f4-cf02-4141-964c-4737b34410ed","Type":"ContainerStarted","Data":"2e4e244cd484f22c1e86bc78baf841e0da8cc68e64f36a21a3b838bf5b9d077f"} Dec 06 03:39:42 crc kubenswrapper[4980]: I1206 03:39:42.421930 4980 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-6bffc96f45-tkm9h" Dec 06 03:39:42 crc kubenswrapper[4980]: I1206 03:39:42.425843 4980 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-6bffc96f45-tkm9h" Dec 06 03:39:42 crc kubenswrapper[4980]: I1206 03:39:42.442881 4980 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-6bffc96f45-tkm9h" podStartSLOduration=99.442856029 podStartE2EDuration="1m39.442856029s" podCreationTimestamp="2025-12-06 03:38:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 03:39:42.43932794 +0000 UTC m=+341.684079221" watchObservedRunningTime="2025-12-06 03:39:42.442856029 +0000 UTC m=+341.687607310" Dec 06 03:39:43 crc kubenswrapper[4980]: I1206 03:39:43.163639 4980 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Dec 06 03:39:44 crc kubenswrapper[4980]: I1206 03:39:44.445195 4980 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Dec 06 03:39:44 crc kubenswrapper[4980]: I1206 03:39:44.585073 4980 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Dec 06 03:39:48 crc kubenswrapper[4980]: I1206 03:39:48.869907 4980 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Dec 06 03:39:49 crc kubenswrapper[4980]: I1206 03:39:49.118738 4980 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-qpchc"] Dec 06 03:39:49 crc kubenswrapper[4980]: I1206 03:39:49.119278 4980 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-879f6c89f-qpchc" podUID="2ac9a45e-ea4a-472e-bd5a-d947a0e2713e" containerName="controller-manager" containerID="cri-o://288edbab674e96c7efebe2535fb2456297f6f828a2250c216e9aef3578076bff" gracePeriod=30 Dec 06 03:39:49 crc kubenswrapper[4980]: I1206 03:39:49.213687 4980 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-2ph5v"] Dec 06 03:39:49 crc kubenswrapper[4980]: I1206 03:39:49.213895 4980 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-2ph5v" podUID="c7db3f86-0e71-46d1-b95c-929069bf8b3a" containerName="route-controller-manager" containerID="cri-o://2ef8b7cf1dcac906424d4e7209c65dd7daa71329776957ae1672c27fedffca30" gracePeriod=30 Dec 06 03:39:49 crc kubenswrapper[4980]: I1206 03:39:49.476778 4980 generic.go:334] "Generic (PLEG): container finished" podID="2ac9a45e-ea4a-472e-bd5a-d947a0e2713e" containerID="288edbab674e96c7efebe2535fb2456297f6f828a2250c216e9aef3578076bff" exitCode=0 Dec 06 03:39:49 crc kubenswrapper[4980]: I1206 03:39:49.476856 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-qpchc" event={"ID":"2ac9a45e-ea4a-472e-bd5a-d947a0e2713e","Type":"ContainerDied","Data":"288edbab674e96c7efebe2535fb2456297f6f828a2250c216e9aef3578076bff"} Dec 06 03:39:49 crc kubenswrapper[4980]: I1206 03:39:49.478399 4980 generic.go:334] "Generic (PLEG): container finished" podID="c7db3f86-0e71-46d1-b95c-929069bf8b3a" containerID="2ef8b7cf1dcac906424d4e7209c65dd7daa71329776957ae1672c27fedffca30" exitCode=0 Dec 06 03:39:49 crc kubenswrapper[4980]: I1206 03:39:49.478445 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-2ph5v" event={"ID":"c7db3f86-0e71-46d1-b95c-929069bf8b3a","Type":"ContainerDied","Data":"2ef8b7cf1dcac906424d4e7209c65dd7daa71329776957ae1672c27fedffca30"} Dec 06 03:39:49 crc kubenswrapper[4980]: I1206 03:39:49.510004 4980 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-qpchc" Dec 06 03:39:49 crc kubenswrapper[4980]: I1206 03:39:49.565342 4980 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-2ph5v" Dec 06 03:39:49 crc kubenswrapper[4980]: I1206 03:39:49.572891 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c7db3f86-0e71-46d1-b95c-929069bf8b3a-serving-cert\") pod \"c7db3f86-0e71-46d1-b95c-929069bf8b3a\" (UID: \"c7db3f86-0e71-46d1-b95c-929069bf8b3a\") " Dec 06 03:39:49 crc kubenswrapper[4980]: I1206 03:39:49.572924 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c7db3f86-0e71-46d1-b95c-929069bf8b3a-config\") pod \"c7db3f86-0e71-46d1-b95c-929069bf8b3a\" (UID: \"c7db3f86-0e71-46d1-b95c-929069bf8b3a\") " Dec 06 03:39:49 crc kubenswrapper[4980]: I1206 03:39:49.572977 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/2ac9a45e-ea4a-472e-bd5a-d947a0e2713e-proxy-ca-bundles\") pod \"2ac9a45e-ea4a-472e-bd5a-d947a0e2713e\" (UID: \"2ac9a45e-ea4a-472e-bd5a-d947a0e2713e\") " Dec 06 03:39:49 crc kubenswrapper[4980]: I1206 03:39:49.572999 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2ac9a45e-ea4a-472e-bd5a-d947a0e2713e-config\") pod \"2ac9a45e-ea4a-472e-bd5a-d947a0e2713e\" (UID: \"2ac9a45e-ea4a-472e-bd5a-d947a0e2713e\") " Dec 06 03:39:49 crc kubenswrapper[4980]: I1206 03:39:49.573027 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n7rjm\" (UniqueName: \"kubernetes.io/projected/c7db3f86-0e71-46d1-b95c-929069bf8b3a-kube-api-access-n7rjm\") pod \"c7db3f86-0e71-46d1-b95c-929069bf8b3a\" (UID: \"c7db3f86-0e71-46d1-b95c-929069bf8b3a\") " Dec 06 03:39:49 crc kubenswrapper[4980]: I1206 03:39:49.573041 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r85l6\" (UniqueName: \"kubernetes.io/projected/2ac9a45e-ea4a-472e-bd5a-d947a0e2713e-kube-api-access-r85l6\") pod \"2ac9a45e-ea4a-472e-bd5a-d947a0e2713e\" (UID: \"2ac9a45e-ea4a-472e-bd5a-d947a0e2713e\") " Dec 06 03:39:49 crc kubenswrapper[4980]: I1206 03:39:49.573063 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2ac9a45e-ea4a-472e-bd5a-d947a0e2713e-serving-cert\") pod \"2ac9a45e-ea4a-472e-bd5a-d947a0e2713e\" (UID: \"2ac9a45e-ea4a-472e-bd5a-d947a0e2713e\") " Dec 06 03:39:49 crc kubenswrapper[4980]: I1206 03:39:49.573086 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/2ac9a45e-ea4a-472e-bd5a-d947a0e2713e-client-ca\") pod \"2ac9a45e-ea4a-472e-bd5a-d947a0e2713e\" (UID: \"2ac9a45e-ea4a-472e-bd5a-d947a0e2713e\") " Dec 06 03:39:49 crc kubenswrapper[4980]: I1206 03:39:49.573106 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/c7db3f86-0e71-46d1-b95c-929069bf8b3a-client-ca\") pod \"c7db3f86-0e71-46d1-b95c-929069bf8b3a\" (UID: \"c7db3f86-0e71-46d1-b95c-929069bf8b3a\") " Dec 06 03:39:49 crc kubenswrapper[4980]: I1206 03:39:49.574501 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c7db3f86-0e71-46d1-b95c-929069bf8b3a-client-ca" (OuterVolumeSpecName: "client-ca") pod "c7db3f86-0e71-46d1-b95c-929069bf8b3a" (UID: "c7db3f86-0e71-46d1-b95c-929069bf8b3a"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 03:39:49 crc kubenswrapper[4980]: I1206 03:39:49.575039 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2ac9a45e-ea4a-472e-bd5a-d947a0e2713e-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "2ac9a45e-ea4a-472e-bd5a-d947a0e2713e" (UID: "2ac9a45e-ea4a-472e-bd5a-d947a0e2713e"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 03:39:49 crc kubenswrapper[4980]: I1206 03:39:49.575718 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c7db3f86-0e71-46d1-b95c-929069bf8b3a-config" (OuterVolumeSpecName: "config") pod "c7db3f86-0e71-46d1-b95c-929069bf8b3a" (UID: "c7db3f86-0e71-46d1-b95c-929069bf8b3a"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 03:39:49 crc kubenswrapper[4980]: I1206 03:39:49.576040 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2ac9a45e-ea4a-472e-bd5a-d947a0e2713e-config" (OuterVolumeSpecName: "config") pod "2ac9a45e-ea4a-472e-bd5a-d947a0e2713e" (UID: "2ac9a45e-ea4a-472e-bd5a-d947a0e2713e"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 03:39:49 crc kubenswrapper[4980]: I1206 03:39:49.576470 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2ac9a45e-ea4a-472e-bd5a-d947a0e2713e-client-ca" (OuterVolumeSpecName: "client-ca") pod "2ac9a45e-ea4a-472e-bd5a-d947a0e2713e" (UID: "2ac9a45e-ea4a-472e-bd5a-d947a0e2713e"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 03:39:49 crc kubenswrapper[4980]: I1206 03:39:49.581560 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2ac9a45e-ea4a-472e-bd5a-d947a0e2713e-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "2ac9a45e-ea4a-472e-bd5a-d947a0e2713e" (UID: "2ac9a45e-ea4a-472e-bd5a-d947a0e2713e"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 03:39:49 crc kubenswrapper[4980]: I1206 03:39:49.581603 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c7db3f86-0e71-46d1-b95c-929069bf8b3a-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "c7db3f86-0e71-46d1-b95c-929069bf8b3a" (UID: "c7db3f86-0e71-46d1-b95c-929069bf8b3a"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 03:39:49 crc kubenswrapper[4980]: I1206 03:39:49.582234 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2ac9a45e-ea4a-472e-bd5a-d947a0e2713e-kube-api-access-r85l6" (OuterVolumeSpecName: "kube-api-access-r85l6") pod "2ac9a45e-ea4a-472e-bd5a-d947a0e2713e" (UID: "2ac9a45e-ea4a-472e-bd5a-d947a0e2713e"). InnerVolumeSpecName "kube-api-access-r85l6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 03:39:49 crc kubenswrapper[4980]: I1206 03:39:49.583206 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c7db3f86-0e71-46d1-b95c-929069bf8b3a-kube-api-access-n7rjm" (OuterVolumeSpecName: "kube-api-access-n7rjm") pod "c7db3f86-0e71-46d1-b95c-929069bf8b3a" (UID: "c7db3f86-0e71-46d1-b95c-929069bf8b3a"). InnerVolumeSpecName "kube-api-access-n7rjm". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 03:39:49 crc kubenswrapper[4980]: I1206 03:39:49.727234 4980 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/2ac9a45e-ea4a-472e-bd5a-d947a0e2713e-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Dec 06 03:39:49 crc kubenswrapper[4980]: I1206 03:39:49.727644 4980 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2ac9a45e-ea4a-472e-bd5a-d947a0e2713e-config\") on node \"crc\" DevicePath \"\"" Dec 06 03:39:49 crc kubenswrapper[4980]: I1206 03:39:49.727749 4980 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r85l6\" (UniqueName: \"kubernetes.io/projected/2ac9a45e-ea4a-472e-bd5a-d947a0e2713e-kube-api-access-r85l6\") on node \"crc\" DevicePath \"\"" Dec 06 03:39:49 crc kubenswrapper[4980]: I1206 03:39:49.727863 4980 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n7rjm\" (UniqueName: \"kubernetes.io/projected/c7db3f86-0e71-46d1-b95c-929069bf8b3a-kube-api-access-n7rjm\") on node \"crc\" DevicePath \"\"" Dec 06 03:39:49 crc kubenswrapper[4980]: I1206 03:39:49.727984 4980 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2ac9a45e-ea4a-472e-bd5a-d947a0e2713e-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 06 03:39:49 crc kubenswrapper[4980]: I1206 03:39:49.728067 4980 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/2ac9a45e-ea4a-472e-bd5a-d947a0e2713e-client-ca\") on node \"crc\" DevicePath \"\"" Dec 06 03:39:49 crc kubenswrapper[4980]: I1206 03:39:49.728145 4980 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/c7db3f86-0e71-46d1-b95c-929069bf8b3a-client-ca\") on node \"crc\" DevicePath \"\"" Dec 06 03:39:49 crc kubenswrapper[4980]: I1206 03:39:49.728232 4980 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c7db3f86-0e71-46d1-b95c-929069bf8b3a-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 06 03:39:49 crc kubenswrapper[4980]: I1206 03:39:49.728313 4980 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c7db3f86-0e71-46d1-b95c-929069bf8b3a-config\") on node \"crc\" DevicePath \"\"" Dec 06 03:39:50 crc kubenswrapper[4980]: I1206 03:39:50.442902 4980 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-7565894c79-s4qkf"] Dec 06 03:39:50 crc kubenswrapper[4980]: E1206 03:39:50.444407 4980 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c7db3f86-0e71-46d1-b95c-929069bf8b3a" containerName="route-controller-manager" Dec 06 03:39:50 crc kubenswrapper[4980]: I1206 03:39:50.444556 4980 state_mem.go:107] "Deleted CPUSet assignment" podUID="c7db3f86-0e71-46d1-b95c-929069bf8b3a" containerName="route-controller-manager" Dec 06 03:39:50 crc kubenswrapper[4980]: E1206 03:39:50.444663 4980 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Dec 06 03:39:50 crc kubenswrapper[4980]: I1206 03:39:50.444738 4980 state_mem.go:107] "Deleted CPUSet assignment" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Dec 06 03:39:50 crc kubenswrapper[4980]: E1206 03:39:50.444820 4980 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2ac9a45e-ea4a-472e-bd5a-d947a0e2713e" containerName="controller-manager" Dec 06 03:39:50 crc kubenswrapper[4980]: I1206 03:39:50.444909 4980 state_mem.go:107] "Deleted CPUSet assignment" podUID="2ac9a45e-ea4a-472e-bd5a-d947a0e2713e" containerName="controller-manager" Dec 06 03:39:50 crc kubenswrapper[4980]: I1206 03:39:50.445154 4980 memory_manager.go:354] "RemoveStaleState removing state" podUID="2ac9a45e-ea4a-472e-bd5a-d947a0e2713e" containerName="controller-manager" Dec 06 03:39:50 crc kubenswrapper[4980]: I1206 03:39:50.445711 4980 memory_manager.go:354] "RemoveStaleState removing state" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Dec 06 03:39:50 crc kubenswrapper[4980]: I1206 03:39:50.445842 4980 memory_manager.go:354] "RemoveStaleState removing state" podUID="c7db3f86-0e71-46d1-b95c-929069bf8b3a" containerName="route-controller-manager" Dec 06 03:39:50 crc kubenswrapper[4980]: I1206 03:39:50.447175 4980 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-55b5d6bbdf-hgqw5"] Dec 06 03:39:50 crc kubenswrapper[4980]: I1206 03:39:50.447355 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-7565894c79-s4qkf" Dec 06 03:39:50 crc kubenswrapper[4980]: I1206 03:39:50.448178 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-55b5d6bbdf-hgqw5" Dec 06 03:39:50 crc kubenswrapper[4980]: I1206 03:39:50.453255 4980 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-7565894c79-s4qkf"] Dec 06 03:39:50 crc kubenswrapper[4980]: I1206 03:39:50.462762 4980 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-55b5d6bbdf-hgqw5"] Dec 06 03:39:50 crc kubenswrapper[4980]: I1206 03:39:50.486496 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-qpchc" event={"ID":"2ac9a45e-ea4a-472e-bd5a-d947a0e2713e","Type":"ContainerDied","Data":"612c0487e2671aaa03994e65c2de3a79a15add54047fd57677387ae7f740077f"} Dec 06 03:39:50 crc kubenswrapper[4980]: I1206 03:39:50.486772 4980 scope.go:117] "RemoveContainer" containerID="288edbab674e96c7efebe2535fb2456297f6f828a2250c216e9aef3578076bff" Dec 06 03:39:50 crc kubenswrapper[4980]: I1206 03:39:50.486781 4980 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-qpchc" Dec 06 03:39:50 crc kubenswrapper[4980]: I1206 03:39:50.489121 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-2ph5v" event={"ID":"c7db3f86-0e71-46d1-b95c-929069bf8b3a","Type":"ContainerDied","Data":"cef8e79f42a30641892997d0344f76c48e1de2b456ff814a4372bab18c9bbabb"} Dec 06 03:39:50 crc kubenswrapper[4980]: I1206 03:39:50.489141 4980 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-2ph5v" Dec 06 03:39:50 crc kubenswrapper[4980]: I1206 03:39:50.503182 4980 scope.go:117] "RemoveContainer" containerID="2ef8b7cf1dcac906424d4e7209c65dd7daa71329776957ae1672c27fedffca30" Dec 06 03:39:50 crc kubenswrapper[4980]: I1206 03:39:50.514734 4980 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-2ph5v"] Dec 06 03:39:50 crc kubenswrapper[4980]: I1206 03:39:50.518874 4980 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-2ph5v"] Dec 06 03:39:50 crc kubenswrapper[4980]: I1206 03:39:50.522380 4980 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-qpchc"] Dec 06 03:39:50 crc kubenswrapper[4980]: I1206 03:39:50.525275 4980 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-qpchc"] Dec 06 03:39:50 crc kubenswrapper[4980]: I1206 03:39:50.538238 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wvk6k\" (UniqueName: \"kubernetes.io/projected/b8cf686e-a11b-43bc-9a8d-d1063f190cde-kube-api-access-wvk6k\") pod \"route-controller-manager-7565894c79-s4qkf\" (UID: \"b8cf686e-a11b-43bc-9a8d-d1063f190cde\") " pod="openshift-route-controller-manager/route-controller-manager-7565894c79-s4qkf" Dec 06 03:39:50 crc kubenswrapper[4980]: I1206 03:39:50.538282 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b9151d5d-626a-4872-b1f3-573394461e88-config\") pod \"controller-manager-55b5d6bbdf-hgqw5\" (UID: \"b9151d5d-626a-4872-b1f3-573394461e88\") " pod="openshift-controller-manager/controller-manager-55b5d6bbdf-hgqw5" Dec 06 03:39:50 crc kubenswrapper[4980]: I1206 03:39:50.538313 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b8cf686e-a11b-43bc-9a8d-d1063f190cde-serving-cert\") pod \"route-controller-manager-7565894c79-s4qkf\" (UID: \"b8cf686e-a11b-43bc-9a8d-d1063f190cde\") " pod="openshift-route-controller-manager/route-controller-manager-7565894c79-s4qkf" Dec 06 03:39:50 crc kubenswrapper[4980]: I1206 03:39:50.538376 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/b8cf686e-a11b-43bc-9a8d-d1063f190cde-client-ca\") pod \"route-controller-manager-7565894c79-s4qkf\" (UID: \"b8cf686e-a11b-43bc-9a8d-d1063f190cde\") " pod="openshift-route-controller-manager/route-controller-manager-7565894c79-s4qkf" Dec 06 03:39:50 crc kubenswrapper[4980]: I1206 03:39:50.538393 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/b9151d5d-626a-4872-b1f3-573394461e88-proxy-ca-bundles\") pod \"controller-manager-55b5d6bbdf-hgqw5\" (UID: \"b9151d5d-626a-4872-b1f3-573394461e88\") " pod="openshift-controller-manager/controller-manager-55b5d6bbdf-hgqw5" Dec 06 03:39:50 crc kubenswrapper[4980]: I1206 03:39:50.538410 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b9151d5d-626a-4872-b1f3-573394461e88-serving-cert\") pod \"controller-manager-55b5d6bbdf-hgqw5\" (UID: \"b9151d5d-626a-4872-b1f3-573394461e88\") " pod="openshift-controller-manager/controller-manager-55b5d6bbdf-hgqw5" Dec 06 03:39:50 crc kubenswrapper[4980]: I1206 03:39:50.538506 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/b9151d5d-626a-4872-b1f3-573394461e88-client-ca\") pod \"controller-manager-55b5d6bbdf-hgqw5\" (UID: \"b9151d5d-626a-4872-b1f3-573394461e88\") " pod="openshift-controller-manager/controller-manager-55b5d6bbdf-hgqw5" Dec 06 03:39:50 crc kubenswrapper[4980]: I1206 03:39:50.538566 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tqrpm\" (UniqueName: \"kubernetes.io/projected/b9151d5d-626a-4872-b1f3-573394461e88-kube-api-access-tqrpm\") pod \"controller-manager-55b5d6bbdf-hgqw5\" (UID: \"b9151d5d-626a-4872-b1f3-573394461e88\") " pod="openshift-controller-manager/controller-manager-55b5d6bbdf-hgqw5" Dec 06 03:39:50 crc kubenswrapper[4980]: I1206 03:39:50.538630 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b8cf686e-a11b-43bc-9a8d-d1063f190cde-config\") pod \"route-controller-manager-7565894c79-s4qkf\" (UID: \"b8cf686e-a11b-43bc-9a8d-d1063f190cde\") " pod="openshift-route-controller-manager/route-controller-manager-7565894c79-s4qkf" Dec 06 03:39:50 crc kubenswrapper[4980]: I1206 03:39:50.640057 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wvk6k\" (UniqueName: \"kubernetes.io/projected/b8cf686e-a11b-43bc-9a8d-d1063f190cde-kube-api-access-wvk6k\") pod \"route-controller-manager-7565894c79-s4qkf\" (UID: \"b8cf686e-a11b-43bc-9a8d-d1063f190cde\") " pod="openshift-route-controller-manager/route-controller-manager-7565894c79-s4qkf" Dec 06 03:39:50 crc kubenswrapper[4980]: I1206 03:39:50.640410 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b9151d5d-626a-4872-b1f3-573394461e88-config\") pod \"controller-manager-55b5d6bbdf-hgqw5\" (UID: \"b9151d5d-626a-4872-b1f3-573394461e88\") " pod="openshift-controller-manager/controller-manager-55b5d6bbdf-hgqw5" Dec 06 03:39:50 crc kubenswrapper[4980]: I1206 03:39:50.640578 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b8cf686e-a11b-43bc-9a8d-d1063f190cde-serving-cert\") pod \"route-controller-manager-7565894c79-s4qkf\" (UID: \"b8cf686e-a11b-43bc-9a8d-d1063f190cde\") " pod="openshift-route-controller-manager/route-controller-manager-7565894c79-s4qkf" Dec 06 03:39:50 crc kubenswrapper[4980]: I1206 03:39:50.640787 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/b8cf686e-a11b-43bc-9a8d-d1063f190cde-client-ca\") pod \"route-controller-manager-7565894c79-s4qkf\" (UID: \"b8cf686e-a11b-43bc-9a8d-d1063f190cde\") " pod="openshift-route-controller-manager/route-controller-manager-7565894c79-s4qkf" Dec 06 03:39:50 crc kubenswrapper[4980]: I1206 03:39:50.640880 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/b9151d5d-626a-4872-b1f3-573394461e88-proxy-ca-bundles\") pod \"controller-manager-55b5d6bbdf-hgqw5\" (UID: \"b9151d5d-626a-4872-b1f3-573394461e88\") " pod="openshift-controller-manager/controller-manager-55b5d6bbdf-hgqw5" Dec 06 03:39:50 crc kubenswrapper[4980]: I1206 03:39:50.640969 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b9151d5d-626a-4872-b1f3-573394461e88-serving-cert\") pod \"controller-manager-55b5d6bbdf-hgqw5\" (UID: \"b9151d5d-626a-4872-b1f3-573394461e88\") " pod="openshift-controller-manager/controller-manager-55b5d6bbdf-hgqw5" Dec 06 03:39:50 crc kubenswrapper[4980]: I1206 03:39:50.641055 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/b9151d5d-626a-4872-b1f3-573394461e88-client-ca\") pod \"controller-manager-55b5d6bbdf-hgqw5\" (UID: \"b9151d5d-626a-4872-b1f3-573394461e88\") " pod="openshift-controller-manager/controller-manager-55b5d6bbdf-hgqw5" Dec 06 03:39:50 crc kubenswrapper[4980]: I1206 03:39:50.641146 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tqrpm\" (UniqueName: \"kubernetes.io/projected/b9151d5d-626a-4872-b1f3-573394461e88-kube-api-access-tqrpm\") pod \"controller-manager-55b5d6bbdf-hgqw5\" (UID: \"b9151d5d-626a-4872-b1f3-573394461e88\") " pod="openshift-controller-manager/controller-manager-55b5d6bbdf-hgqw5" Dec 06 03:39:50 crc kubenswrapper[4980]: I1206 03:39:50.641250 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b8cf686e-a11b-43bc-9a8d-d1063f190cde-config\") pod \"route-controller-manager-7565894c79-s4qkf\" (UID: \"b8cf686e-a11b-43bc-9a8d-d1063f190cde\") " pod="openshift-route-controller-manager/route-controller-manager-7565894c79-s4qkf" Dec 06 03:39:50 crc kubenswrapper[4980]: I1206 03:39:50.641942 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/b9151d5d-626a-4872-b1f3-573394461e88-client-ca\") pod \"controller-manager-55b5d6bbdf-hgqw5\" (UID: \"b9151d5d-626a-4872-b1f3-573394461e88\") " pod="openshift-controller-manager/controller-manager-55b5d6bbdf-hgqw5" Dec 06 03:39:50 crc kubenswrapper[4980]: I1206 03:39:50.641997 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/b8cf686e-a11b-43bc-9a8d-d1063f190cde-client-ca\") pod \"route-controller-manager-7565894c79-s4qkf\" (UID: \"b8cf686e-a11b-43bc-9a8d-d1063f190cde\") " pod="openshift-route-controller-manager/route-controller-manager-7565894c79-s4qkf" Dec 06 03:39:50 crc kubenswrapper[4980]: I1206 03:39:50.642105 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b9151d5d-626a-4872-b1f3-573394461e88-config\") pod \"controller-manager-55b5d6bbdf-hgqw5\" (UID: \"b9151d5d-626a-4872-b1f3-573394461e88\") " pod="openshift-controller-manager/controller-manager-55b5d6bbdf-hgqw5" Dec 06 03:39:50 crc kubenswrapper[4980]: I1206 03:39:50.642196 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/b9151d5d-626a-4872-b1f3-573394461e88-proxy-ca-bundles\") pod \"controller-manager-55b5d6bbdf-hgqw5\" (UID: \"b9151d5d-626a-4872-b1f3-573394461e88\") " pod="openshift-controller-manager/controller-manager-55b5d6bbdf-hgqw5" Dec 06 03:39:50 crc kubenswrapper[4980]: I1206 03:39:50.642801 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b8cf686e-a11b-43bc-9a8d-d1063f190cde-config\") pod \"route-controller-manager-7565894c79-s4qkf\" (UID: \"b8cf686e-a11b-43bc-9a8d-d1063f190cde\") " pod="openshift-route-controller-manager/route-controller-manager-7565894c79-s4qkf" Dec 06 03:39:50 crc kubenswrapper[4980]: I1206 03:39:50.645264 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b8cf686e-a11b-43bc-9a8d-d1063f190cde-serving-cert\") pod \"route-controller-manager-7565894c79-s4qkf\" (UID: \"b8cf686e-a11b-43bc-9a8d-d1063f190cde\") " pod="openshift-route-controller-manager/route-controller-manager-7565894c79-s4qkf" Dec 06 03:39:50 crc kubenswrapper[4980]: I1206 03:39:50.655005 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b9151d5d-626a-4872-b1f3-573394461e88-serving-cert\") pod \"controller-manager-55b5d6bbdf-hgqw5\" (UID: \"b9151d5d-626a-4872-b1f3-573394461e88\") " pod="openshift-controller-manager/controller-manager-55b5d6bbdf-hgqw5" Dec 06 03:39:50 crc kubenswrapper[4980]: I1206 03:39:50.655589 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tqrpm\" (UniqueName: \"kubernetes.io/projected/b9151d5d-626a-4872-b1f3-573394461e88-kube-api-access-tqrpm\") pod \"controller-manager-55b5d6bbdf-hgqw5\" (UID: \"b9151d5d-626a-4872-b1f3-573394461e88\") " pod="openshift-controller-manager/controller-manager-55b5d6bbdf-hgqw5" Dec 06 03:39:50 crc kubenswrapper[4980]: I1206 03:39:50.661453 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wvk6k\" (UniqueName: \"kubernetes.io/projected/b8cf686e-a11b-43bc-9a8d-d1063f190cde-kube-api-access-wvk6k\") pod \"route-controller-manager-7565894c79-s4qkf\" (UID: \"b8cf686e-a11b-43bc-9a8d-d1063f190cde\") " pod="openshift-route-controller-manager/route-controller-manager-7565894c79-s4qkf" Dec 06 03:39:50 crc kubenswrapper[4980]: I1206 03:39:50.776257 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-7565894c79-s4qkf" Dec 06 03:39:50 crc kubenswrapper[4980]: I1206 03:39:50.786048 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-55b5d6bbdf-hgqw5" Dec 06 03:39:50 crc kubenswrapper[4980]: I1206 03:39:50.959083 4980 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-7565894c79-s4qkf"] Dec 06 03:39:51 crc kubenswrapper[4980]: I1206 03:39:51.001535 4980 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-55b5d6bbdf-hgqw5"] Dec 06 03:39:51 crc kubenswrapper[4980]: W1206 03:39:51.005213 4980 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb9151d5d_626a_4872_b1f3_573394461e88.slice/crio-379e5949efe35e00227115a5aa4d78fa6e6e973eb0be0fe0db2b92c8c7c24035 WatchSource:0}: Error finding container 379e5949efe35e00227115a5aa4d78fa6e6e973eb0be0fe0db2b92c8c7c24035: Status 404 returned error can't find the container with id 379e5949efe35e00227115a5aa4d78fa6e6e973eb0be0fe0db2b92c8c7c24035 Dec 06 03:39:51 crc kubenswrapper[4980]: I1206 03:39:51.192409 4980 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2ac9a45e-ea4a-472e-bd5a-d947a0e2713e" path="/var/lib/kubelet/pods/2ac9a45e-ea4a-472e-bd5a-d947a0e2713e/volumes" Dec 06 03:39:51 crc kubenswrapper[4980]: I1206 03:39:51.193308 4980 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c7db3f86-0e71-46d1-b95c-929069bf8b3a" path="/var/lib/kubelet/pods/c7db3f86-0e71-46d1-b95c-929069bf8b3a/volumes" Dec 06 03:39:51 crc kubenswrapper[4980]: I1206 03:39:51.500606 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-7565894c79-s4qkf" event={"ID":"b8cf686e-a11b-43bc-9a8d-d1063f190cde","Type":"ContainerStarted","Data":"b9a76f743fd070e3ef8a67789efbf812ec8eb2952d3e6f2461272a02439febdd"} Dec 06 03:39:51 crc kubenswrapper[4980]: I1206 03:39:51.500980 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-7565894c79-s4qkf" event={"ID":"b8cf686e-a11b-43bc-9a8d-d1063f190cde","Type":"ContainerStarted","Data":"f1bec8bb0745a38310b7ef4b051c98a5111ecabbfb3d2c4037408d5408fd5282"} Dec 06 03:39:51 crc kubenswrapper[4980]: I1206 03:39:51.501453 4980 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-7565894c79-s4qkf" Dec 06 03:39:51 crc kubenswrapper[4980]: I1206 03:39:51.502992 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-55b5d6bbdf-hgqw5" event={"ID":"b9151d5d-626a-4872-b1f3-573394461e88","Type":"ContainerStarted","Data":"5193c1f55e084ebd1d67fca85e2644d601e2c7c1bf3e27ca641b00b0b502e2d8"} Dec 06 03:39:51 crc kubenswrapper[4980]: I1206 03:39:51.503041 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-55b5d6bbdf-hgqw5" event={"ID":"b9151d5d-626a-4872-b1f3-573394461e88","Type":"ContainerStarted","Data":"379e5949efe35e00227115a5aa4d78fa6e6e973eb0be0fe0db2b92c8c7c24035"} Dec 06 03:39:51 crc kubenswrapper[4980]: I1206 03:39:51.503251 4980 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-55b5d6bbdf-hgqw5" Dec 06 03:39:51 crc kubenswrapper[4980]: I1206 03:39:51.508102 4980 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-55b5d6bbdf-hgqw5" Dec 06 03:39:51 crc kubenswrapper[4980]: I1206 03:39:51.519048 4980 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-7565894c79-s4qkf" podStartSLOduration=2.519033621 podStartE2EDuration="2.519033621s" podCreationTimestamp="2025-12-06 03:39:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 03:39:51.515407929 +0000 UTC m=+350.760159200" watchObservedRunningTime="2025-12-06 03:39:51.519033621 +0000 UTC m=+350.763784902" Dec 06 03:39:51 crc kubenswrapper[4980]: I1206 03:39:51.530620 4980 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-55b5d6bbdf-hgqw5" podStartSLOduration=2.530600598 podStartE2EDuration="2.530600598s" podCreationTimestamp="2025-12-06 03:39:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 03:39:51.530395502 +0000 UTC m=+350.775146783" watchObservedRunningTime="2025-12-06 03:39:51.530600598 +0000 UTC m=+350.775351879" Dec 06 03:39:51 crc kubenswrapper[4980]: I1206 03:39:51.741948 4980 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-7565894c79-s4qkf" Dec 06 03:39:55 crc kubenswrapper[4980]: I1206 03:39:55.137349 4980 patch_prober.go:28] interesting pod/machine-config-daemon-r5zfb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 03:39:55 crc kubenswrapper[4980]: I1206 03:39:55.138098 4980 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-r5zfb" podUID="320f44d1-a671-4a91-b328-a8b0fdd8f23a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 03:40:18 crc kubenswrapper[4980]: I1206 03:40:18.396743 4980 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-55b5d6bbdf-hgqw5"] Dec 06 03:40:18 crc kubenswrapper[4980]: I1206 03:40:18.399360 4980 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-55b5d6bbdf-hgqw5" podUID="b9151d5d-626a-4872-b1f3-573394461e88" containerName="controller-manager" containerID="cri-o://5193c1f55e084ebd1d67fca85e2644d601e2c7c1bf3e27ca641b00b0b502e2d8" gracePeriod=30 Dec 06 03:40:18 crc kubenswrapper[4980]: I1206 03:40:18.404195 4980 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-7565894c79-s4qkf"] Dec 06 03:40:18 crc kubenswrapper[4980]: I1206 03:40:18.404507 4980 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-7565894c79-s4qkf" podUID="b8cf686e-a11b-43bc-9a8d-d1063f190cde" containerName="route-controller-manager" containerID="cri-o://b9a76f743fd070e3ef8a67789efbf812ec8eb2952d3e6f2461272a02439febdd" gracePeriod=30 Dec 06 03:40:18 crc kubenswrapper[4980]: I1206 03:40:18.691893 4980 generic.go:334] "Generic (PLEG): container finished" podID="b8cf686e-a11b-43bc-9a8d-d1063f190cde" containerID="b9a76f743fd070e3ef8a67789efbf812ec8eb2952d3e6f2461272a02439febdd" exitCode=0 Dec 06 03:40:18 crc kubenswrapper[4980]: I1206 03:40:18.691953 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-7565894c79-s4qkf" event={"ID":"b8cf686e-a11b-43bc-9a8d-d1063f190cde","Type":"ContainerDied","Data":"b9a76f743fd070e3ef8a67789efbf812ec8eb2952d3e6f2461272a02439febdd"} Dec 06 03:40:18 crc kubenswrapper[4980]: I1206 03:40:18.694456 4980 generic.go:334] "Generic (PLEG): container finished" podID="b9151d5d-626a-4872-b1f3-573394461e88" containerID="5193c1f55e084ebd1d67fca85e2644d601e2c7c1bf3e27ca641b00b0b502e2d8" exitCode=0 Dec 06 03:40:18 crc kubenswrapper[4980]: I1206 03:40:18.694567 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-55b5d6bbdf-hgqw5" event={"ID":"b9151d5d-626a-4872-b1f3-573394461e88","Type":"ContainerDied","Data":"5193c1f55e084ebd1d67fca85e2644d601e2c7c1bf3e27ca641b00b0b502e2d8"} Dec 06 03:40:18 crc kubenswrapper[4980]: I1206 03:40:18.965307 4980 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-7565894c79-s4qkf" Dec 06 03:40:18 crc kubenswrapper[4980]: I1206 03:40:18.968405 4980 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-55b5d6bbdf-hgqw5" Dec 06 03:40:19 crc kubenswrapper[4980]: I1206 03:40:19.118949 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/b9151d5d-626a-4872-b1f3-573394461e88-proxy-ca-bundles\") pod \"b9151d5d-626a-4872-b1f3-573394461e88\" (UID: \"b9151d5d-626a-4872-b1f3-573394461e88\") " Dec 06 03:40:19 crc kubenswrapper[4980]: I1206 03:40:19.119012 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wvk6k\" (UniqueName: \"kubernetes.io/projected/b8cf686e-a11b-43bc-9a8d-d1063f190cde-kube-api-access-wvk6k\") pod \"b8cf686e-a11b-43bc-9a8d-d1063f190cde\" (UID: \"b8cf686e-a11b-43bc-9a8d-d1063f190cde\") " Dec 06 03:40:19 crc kubenswrapper[4980]: I1206 03:40:19.119039 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tqrpm\" (UniqueName: \"kubernetes.io/projected/b9151d5d-626a-4872-b1f3-573394461e88-kube-api-access-tqrpm\") pod \"b9151d5d-626a-4872-b1f3-573394461e88\" (UID: \"b9151d5d-626a-4872-b1f3-573394461e88\") " Dec 06 03:40:19 crc kubenswrapper[4980]: I1206 03:40:19.119065 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/b9151d5d-626a-4872-b1f3-573394461e88-client-ca\") pod \"b9151d5d-626a-4872-b1f3-573394461e88\" (UID: \"b9151d5d-626a-4872-b1f3-573394461e88\") " Dec 06 03:40:19 crc kubenswrapper[4980]: I1206 03:40:19.119121 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/b8cf686e-a11b-43bc-9a8d-d1063f190cde-client-ca\") pod \"b8cf686e-a11b-43bc-9a8d-d1063f190cde\" (UID: \"b8cf686e-a11b-43bc-9a8d-d1063f190cde\") " Dec 06 03:40:19 crc kubenswrapper[4980]: I1206 03:40:19.119160 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b8cf686e-a11b-43bc-9a8d-d1063f190cde-serving-cert\") pod \"b8cf686e-a11b-43bc-9a8d-d1063f190cde\" (UID: \"b8cf686e-a11b-43bc-9a8d-d1063f190cde\") " Dec 06 03:40:19 crc kubenswrapper[4980]: I1206 03:40:19.119191 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b9151d5d-626a-4872-b1f3-573394461e88-serving-cert\") pod \"b9151d5d-626a-4872-b1f3-573394461e88\" (UID: \"b9151d5d-626a-4872-b1f3-573394461e88\") " Dec 06 03:40:19 crc kubenswrapper[4980]: I1206 03:40:19.119227 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b9151d5d-626a-4872-b1f3-573394461e88-config\") pod \"b9151d5d-626a-4872-b1f3-573394461e88\" (UID: \"b9151d5d-626a-4872-b1f3-573394461e88\") " Dec 06 03:40:19 crc kubenswrapper[4980]: I1206 03:40:19.119262 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b8cf686e-a11b-43bc-9a8d-d1063f190cde-config\") pod \"b8cf686e-a11b-43bc-9a8d-d1063f190cde\" (UID: \"b8cf686e-a11b-43bc-9a8d-d1063f190cde\") " Dec 06 03:40:19 crc kubenswrapper[4980]: I1206 03:40:19.121015 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b8cf686e-a11b-43bc-9a8d-d1063f190cde-config" (OuterVolumeSpecName: "config") pod "b8cf686e-a11b-43bc-9a8d-d1063f190cde" (UID: "b8cf686e-a11b-43bc-9a8d-d1063f190cde"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 03:40:19 crc kubenswrapper[4980]: I1206 03:40:19.121758 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b8cf686e-a11b-43bc-9a8d-d1063f190cde-client-ca" (OuterVolumeSpecName: "client-ca") pod "b8cf686e-a11b-43bc-9a8d-d1063f190cde" (UID: "b8cf686e-a11b-43bc-9a8d-d1063f190cde"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 03:40:19 crc kubenswrapper[4980]: I1206 03:40:19.126854 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b8cf686e-a11b-43bc-9a8d-d1063f190cde-kube-api-access-wvk6k" (OuterVolumeSpecName: "kube-api-access-wvk6k") pod "b8cf686e-a11b-43bc-9a8d-d1063f190cde" (UID: "b8cf686e-a11b-43bc-9a8d-d1063f190cde"). InnerVolumeSpecName "kube-api-access-wvk6k". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 03:40:19 crc kubenswrapper[4980]: I1206 03:40:19.126857 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b9151d5d-626a-4872-b1f3-573394461e88-client-ca" (OuterVolumeSpecName: "client-ca") pod "b9151d5d-626a-4872-b1f3-573394461e88" (UID: "b9151d5d-626a-4872-b1f3-573394461e88"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 03:40:19 crc kubenswrapper[4980]: I1206 03:40:19.127281 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b9151d5d-626a-4872-b1f3-573394461e88-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "b9151d5d-626a-4872-b1f3-573394461e88" (UID: "b9151d5d-626a-4872-b1f3-573394461e88"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 03:40:19 crc kubenswrapper[4980]: I1206 03:40:19.129899 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b9151d5d-626a-4872-b1f3-573394461e88-config" (OuterVolumeSpecName: "config") pod "b9151d5d-626a-4872-b1f3-573394461e88" (UID: "b9151d5d-626a-4872-b1f3-573394461e88"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 03:40:19 crc kubenswrapper[4980]: I1206 03:40:19.130028 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b9151d5d-626a-4872-b1f3-573394461e88-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "b9151d5d-626a-4872-b1f3-573394461e88" (UID: "b9151d5d-626a-4872-b1f3-573394461e88"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 03:40:19 crc kubenswrapper[4980]: I1206 03:40:19.141332 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b8cf686e-a11b-43bc-9a8d-d1063f190cde-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "b8cf686e-a11b-43bc-9a8d-d1063f190cde" (UID: "b8cf686e-a11b-43bc-9a8d-d1063f190cde"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 03:40:19 crc kubenswrapper[4980]: I1206 03:40:19.143161 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b9151d5d-626a-4872-b1f3-573394461e88-kube-api-access-tqrpm" (OuterVolumeSpecName: "kube-api-access-tqrpm") pod "b9151d5d-626a-4872-b1f3-573394461e88" (UID: "b9151d5d-626a-4872-b1f3-573394461e88"). InnerVolumeSpecName "kube-api-access-tqrpm". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 03:40:19 crc kubenswrapper[4980]: I1206 03:40:19.220304 4980 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b8cf686e-a11b-43bc-9a8d-d1063f190cde-config\") on node \"crc\" DevicePath \"\"" Dec 06 03:40:19 crc kubenswrapper[4980]: I1206 03:40:19.220341 4980 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/b9151d5d-626a-4872-b1f3-573394461e88-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Dec 06 03:40:19 crc kubenswrapper[4980]: I1206 03:40:19.220351 4980 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wvk6k\" (UniqueName: \"kubernetes.io/projected/b8cf686e-a11b-43bc-9a8d-d1063f190cde-kube-api-access-wvk6k\") on node \"crc\" DevicePath \"\"" Dec 06 03:40:19 crc kubenswrapper[4980]: I1206 03:40:19.220362 4980 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tqrpm\" (UniqueName: \"kubernetes.io/projected/b9151d5d-626a-4872-b1f3-573394461e88-kube-api-access-tqrpm\") on node \"crc\" DevicePath \"\"" Dec 06 03:40:19 crc kubenswrapper[4980]: I1206 03:40:19.220374 4980 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/b9151d5d-626a-4872-b1f3-573394461e88-client-ca\") on node \"crc\" DevicePath \"\"" Dec 06 03:40:19 crc kubenswrapper[4980]: I1206 03:40:19.220385 4980 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/b8cf686e-a11b-43bc-9a8d-d1063f190cde-client-ca\") on node \"crc\" DevicePath \"\"" Dec 06 03:40:19 crc kubenswrapper[4980]: I1206 03:40:19.220395 4980 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b8cf686e-a11b-43bc-9a8d-d1063f190cde-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 06 03:40:19 crc kubenswrapper[4980]: I1206 03:40:19.220407 4980 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b9151d5d-626a-4872-b1f3-573394461e88-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 06 03:40:19 crc kubenswrapper[4980]: I1206 03:40:19.220417 4980 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b9151d5d-626a-4872-b1f3-573394461e88-config\") on node \"crc\" DevicePath \"\"" Dec 06 03:40:19 crc kubenswrapper[4980]: I1206 03:40:19.409890 4980 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-55b5d6bbdf-v9mg9"] Dec 06 03:40:19 crc kubenswrapper[4980]: E1206 03:40:19.410195 4980 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b8cf686e-a11b-43bc-9a8d-d1063f190cde" containerName="route-controller-manager" Dec 06 03:40:19 crc kubenswrapper[4980]: I1206 03:40:19.410213 4980 state_mem.go:107] "Deleted CPUSet assignment" podUID="b8cf686e-a11b-43bc-9a8d-d1063f190cde" containerName="route-controller-manager" Dec 06 03:40:19 crc kubenswrapper[4980]: E1206 03:40:19.410233 4980 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b9151d5d-626a-4872-b1f3-573394461e88" containerName="controller-manager" Dec 06 03:40:19 crc kubenswrapper[4980]: I1206 03:40:19.410241 4980 state_mem.go:107] "Deleted CPUSet assignment" podUID="b9151d5d-626a-4872-b1f3-573394461e88" containerName="controller-manager" Dec 06 03:40:19 crc kubenswrapper[4980]: I1206 03:40:19.410435 4980 memory_manager.go:354] "RemoveStaleState removing state" podUID="b8cf686e-a11b-43bc-9a8d-d1063f190cde" containerName="route-controller-manager" Dec 06 03:40:19 crc kubenswrapper[4980]: I1206 03:40:19.410452 4980 memory_manager.go:354] "RemoveStaleState removing state" podUID="b9151d5d-626a-4872-b1f3-573394461e88" containerName="controller-manager" Dec 06 03:40:19 crc kubenswrapper[4980]: I1206 03:40:19.410968 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-55b5d6bbdf-v9mg9" Dec 06 03:40:19 crc kubenswrapper[4980]: I1206 03:40:19.423160 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8ch5w\" (UniqueName: \"kubernetes.io/projected/aff033cf-643b-4060-96c2-9d65b9ceb724-kube-api-access-8ch5w\") pod \"controller-manager-55b5d6bbdf-v9mg9\" (UID: \"aff033cf-643b-4060-96c2-9d65b9ceb724\") " pod="openshift-controller-manager/controller-manager-55b5d6bbdf-v9mg9" Dec 06 03:40:19 crc kubenswrapper[4980]: I1206 03:40:19.423201 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/aff033cf-643b-4060-96c2-9d65b9ceb724-client-ca\") pod \"controller-manager-55b5d6bbdf-v9mg9\" (UID: \"aff033cf-643b-4060-96c2-9d65b9ceb724\") " pod="openshift-controller-manager/controller-manager-55b5d6bbdf-v9mg9" Dec 06 03:40:19 crc kubenswrapper[4980]: I1206 03:40:19.423240 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/aff033cf-643b-4060-96c2-9d65b9ceb724-config\") pod \"controller-manager-55b5d6bbdf-v9mg9\" (UID: \"aff033cf-643b-4060-96c2-9d65b9ceb724\") " pod="openshift-controller-manager/controller-manager-55b5d6bbdf-v9mg9" Dec 06 03:40:19 crc kubenswrapper[4980]: I1206 03:40:19.423342 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/aff033cf-643b-4060-96c2-9d65b9ceb724-proxy-ca-bundles\") pod \"controller-manager-55b5d6bbdf-v9mg9\" (UID: \"aff033cf-643b-4060-96c2-9d65b9ceb724\") " pod="openshift-controller-manager/controller-manager-55b5d6bbdf-v9mg9" Dec 06 03:40:19 crc kubenswrapper[4980]: I1206 03:40:19.423359 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/aff033cf-643b-4060-96c2-9d65b9ceb724-serving-cert\") pod \"controller-manager-55b5d6bbdf-v9mg9\" (UID: \"aff033cf-643b-4060-96c2-9d65b9ceb724\") " pod="openshift-controller-manager/controller-manager-55b5d6bbdf-v9mg9" Dec 06 03:40:19 crc kubenswrapper[4980]: I1206 03:40:19.428125 4980 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-55b5d6bbdf-v9mg9"] Dec 06 03:40:19 crc kubenswrapper[4980]: I1206 03:40:19.464656 4980 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-677b664cd9-cbb88"] Dec 06 03:40:19 crc kubenswrapper[4980]: I1206 03:40:19.465304 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-677b664cd9-cbb88" Dec 06 03:40:19 crc kubenswrapper[4980]: I1206 03:40:19.474712 4980 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-677b664cd9-cbb88"] Dec 06 03:40:19 crc kubenswrapper[4980]: I1206 03:40:19.523975 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7mrl5\" (UniqueName: \"kubernetes.io/projected/c515dcbb-b1bb-49c4-9ea2-fc80d5fc3d9e-kube-api-access-7mrl5\") pod \"route-controller-manager-677b664cd9-cbb88\" (UID: \"c515dcbb-b1bb-49c4-9ea2-fc80d5fc3d9e\") " pod="openshift-route-controller-manager/route-controller-manager-677b664cd9-cbb88" Dec 06 03:40:19 crc kubenswrapper[4980]: I1206 03:40:19.524058 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/aff033cf-643b-4060-96c2-9d65b9ceb724-proxy-ca-bundles\") pod \"controller-manager-55b5d6bbdf-v9mg9\" (UID: \"aff033cf-643b-4060-96c2-9d65b9ceb724\") " pod="openshift-controller-manager/controller-manager-55b5d6bbdf-v9mg9" Dec 06 03:40:19 crc kubenswrapper[4980]: I1206 03:40:19.524086 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/aff033cf-643b-4060-96c2-9d65b9ceb724-serving-cert\") pod \"controller-manager-55b5d6bbdf-v9mg9\" (UID: \"aff033cf-643b-4060-96c2-9d65b9ceb724\") " pod="openshift-controller-manager/controller-manager-55b5d6bbdf-v9mg9" Dec 06 03:40:19 crc kubenswrapper[4980]: I1206 03:40:19.524149 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/c515dcbb-b1bb-49c4-9ea2-fc80d5fc3d9e-client-ca\") pod \"route-controller-manager-677b664cd9-cbb88\" (UID: \"c515dcbb-b1bb-49c4-9ea2-fc80d5fc3d9e\") " pod="openshift-route-controller-manager/route-controller-manager-677b664cd9-cbb88" Dec 06 03:40:19 crc kubenswrapper[4980]: I1206 03:40:19.524216 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8ch5w\" (UniqueName: \"kubernetes.io/projected/aff033cf-643b-4060-96c2-9d65b9ceb724-kube-api-access-8ch5w\") pod \"controller-manager-55b5d6bbdf-v9mg9\" (UID: \"aff033cf-643b-4060-96c2-9d65b9ceb724\") " pod="openshift-controller-manager/controller-manager-55b5d6bbdf-v9mg9" Dec 06 03:40:19 crc kubenswrapper[4980]: I1206 03:40:19.524238 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/aff033cf-643b-4060-96c2-9d65b9ceb724-client-ca\") pod \"controller-manager-55b5d6bbdf-v9mg9\" (UID: \"aff033cf-643b-4060-96c2-9d65b9ceb724\") " pod="openshift-controller-manager/controller-manager-55b5d6bbdf-v9mg9" Dec 06 03:40:19 crc kubenswrapper[4980]: I1206 03:40:19.524264 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c515dcbb-b1bb-49c4-9ea2-fc80d5fc3d9e-config\") pod \"route-controller-manager-677b664cd9-cbb88\" (UID: \"c515dcbb-b1bb-49c4-9ea2-fc80d5fc3d9e\") " pod="openshift-route-controller-manager/route-controller-manager-677b664cd9-cbb88" Dec 06 03:40:19 crc kubenswrapper[4980]: I1206 03:40:19.524287 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/aff033cf-643b-4060-96c2-9d65b9ceb724-config\") pod \"controller-manager-55b5d6bbdf-v9mg9\" (UID: \"aff033cf-643b-4060-96c2-9d65b9ceb724\") " pod="openshift-controller-manager/controller-manager-55b5d6bbdf-v9mg9" Dec 06 03:40:19 crc kubenswrapper[4980]: I1206 03:40:19.524330 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c515dcbb-b1bb-49c4-9ea2-fc80d5fc3d9e-serving-cert\") pod \"route-controller-manager-677b664cd9-cbb88\" (UID: \"c515dcbb-b1bb-49c4-9ea2-fc80d5fc3d9e\") " pod="openshift-route-controller-manager/route-controller-manager-677b664cd9-cbb88" Dec 06 03:40:19 crc kubenswrapper[4980]: I1206 03:40:19.525430 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/aff033cf-643b-4060-96c2-9d65b9ceb724-client-ca\") pod \"controller-manager-55b5d6bbdf-v9mg9\" (UID: \"aff033cf-643b-4060-96c2-9d65b9ceb724\") " pod="openshift-controller-manager/controller-manager-55b5d6bbdf-v9mg9" Dec 06 03:40:19 crc kubenswrapper[4980]: I1206 03:40:19.525868 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/aff033cf-643b-4060-96c2-9d65b9ceb724-config\") pod \"controller-manager-55b5d6bbdf-v9mg9\" (UID: \"aff033cf-643b-4060-96c2-9d65b9ceb724\") " pod="openshift-controller-manager/controller-manager-55b5d6bbdf-v9mg9" Dec 06 03:40:19 crc kubenswrapper[4980]: I1206 03:40:19.526539 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/aff033cf-643b-4060-96c2-9d65b9ceb724-proxy-ca-bundles\") pod \"controller-manager-55b5d6bbdf-v9mg9\" (UID: \"aff033cf-643b-4060-96c2-9d65b9ceb724\") " pod="openshift-controller-manager/controller-manager-55b5d6bbdf-v9mg9" Dec 06 03:40:19 crc kubenswrapper[4980]: I1206 03:40:19.537001 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/aff033cf-643b-4060-96c2-9d65b9ceb724-serving-cert\") pod \"controller-manager-55b5d6bbdf-v9mg9\" (UID: \"aff033cf-643b-4060-96c2-9d65b9ceb724\") " pod="openshift-controller-manager/controller-manager-55b5d6bbdf-v9mg9" Dec 06 03:40:19 crc kubenswrapper[4980]: I1206 03:40:19.538873 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8ch5w\" (UniqueName: \"kubernetes.io/projected/aff033cf-643b-4060-96c2-9d65b9ceb724-kube-api-access-8ch5w\") pod \"controller-manager-55b5d6bbdf-v9mg9\" (UID: \"aff033cf-643b-4060-96c2-9d65b9ceb724\") " pod="openshift-controller-manager/controller-manager-55b5d6bbdf-v9mg9" Dec 06 03:40:19 crc kubenswrapper[4980]: I1206 03:40:19.625147 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c515dcbb-b1bb-49c4-9ea2-fc80d5fc3d9e-serving-cert\") pod \"route-controller-manager-677b664cd9-cbb88\" (UID: \"c515dcbb-b1bb-49c4-9ea2-fc80d5fc3d9e\") " pod="openshift-route-controller-manager/route-controller-manager-677b664cd9-cbb88" Dec 06 03:40:19 crc kubenswrapper[4980]: I1206 03:40:19.625211 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7mrl5\" (UniqueName: \"kubernetes.io/projected/c515dcbb-b1bb-49c4-9ea2-fc80d5fc3d9e-kube-api-access-7mrl5\") pod \"route-controller-manager-677b664cd9-cbb88\" (UID: \"c515dcbb-b1bb-49c4-9ea2-fc80d5fc3d9e\") " pod="openshift-route-controller-manager/route-controller-manager-677b664cd9-cbb88" Dec 06 03:40:19 crc kubenswrapper[4980]: I1206 03:40:19.625244 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/c515dcbb-b1bb-49c4-9ea2-fc80d5fc3d9e-client-ca\") pod \"route-controller-manager-677b664cd9-cbb88\" (UID: \"c515dcbb-b1bb-49c4-9ea2-fc80d5fc3d9e\") " pod="openshift-route-controller-manager/route-controller-manager-677b664cd9-cbb88" Dec 06 03:40:19 crc kubenswrapper[4980]: I1206 03:40:19.625281 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c515dcbb-b1bb-49c4-9ea2-fc80d5fc3d9e-config\") pod \"route-controller-manager-677b664cd9-cbb88\" (UID: \"c515dcbb-b1bb-49c4-9ea2-fc80d5fc3d9e\") " pod="openshift-route-controller-manager/route-controller-manager-677b664cd9-cbb88" Dec 06 03:40:19 crc kubenswrapper[4980]: I1206 03:40:19.626339 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c515dcbb-b1bb-49c4-9ea2-fc80d5fc3d9e-config\") pod \"route-controller-manager-677b664cd9-cbb88\" (UID: \"c515dcbb-b1bb-49c4-9ea2-fc80d5fc3d9e\") " pod="openshift-route-controller-manager/route-controller-manager-677b664cd9-cbb88" Dec 06 03:40:19 crc kubenswrapper[4980]: I1206 03:40:19.628000 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/c515dcbb-b1bb-49c4-9ea2-fc80d5fc3d9e-client-ca\") pod \"route-controller-manager-677b664cd9-cbb88\" (UID: \"c515dcbb-b1bb-49c4-9ea2-fc80d5fc3d9e\") " pod="openshift-route-controller-manager/route-controller-manager-677b664cd9-cbb88" Dec 06 03:40:19 crc kubenswrapper[4980]: I1206 03:40:19.632955 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c515dcbb-b1bb-49c4-9ea2-fc80d5fc3d9e-serving-cert\") pod \"route-controller-manager-677b664cd9-cbb88\" (UID: \"c515dcbb-b1bb-49c4-9ea2-fc80d5fc3d9e\") " pod="openshift-route-controller-manager/route-controller-manager-677b664cd9-cbb88" Dec 06 03:40:19 crc kubenswrapper[4980]: I1206 03:40:19.645080 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7mrl5\" (UniqueName: \"kubernetes.io/projected/c515dcbb-b1bb-49c4-9ea2-fc80d5fc3d9e-kube-api-access-7mrl5\") pod \"route-controller-manager-677b664cd9-cbb88\" (UID: \"c515dcbb-b1bb-49c4-9ea2-fc80d5fc3d9e\") " pod="openshift-route-controller-manager/route-controller-manager-677b664cd9-cbb88" Dec 06 03:40:19 crc kubenswrapper[4980]: I1206 03:40:19.703900 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-7565894c79-s4qkf" event={"ID":"b8cf686e-a11b-43bc-9a8d-d1063f190cde","Type":"ContainerDied","Data":"f1bec8bb0745a38310b7ef4b051c98a5111ecabbfb3d2c4037408d5408fd5282"} Dec 06 03:40:19 crc kubenswrapper[4980]: I1206 03:40:19.704040 4980 scope.go:117] "RemoveContainer" containerID="b9a76f743fd070e3ef8a67789efbf812ec8eb2952d3e6f2461272a02439febdd" Dec 06 03:40:19 crc kubenswrapper[4980]: I1206 03:40:19.704239 4980 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-7565894c79-s4qkf" Dec 06 03:40:19 crc kubenswrapper[4980]: I1206 03:40:19.709967 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-55b5d6bbdf-hgqw5" event={"ID":"b9151d5d-626a-4872-b1f3-573394461e88","Type":"ContainerDied","Data":"379e5949efe35e00227115a5aa4d78fa6e6e973eb0be0fe0db2b92c8c7c24035"} Dec 06 03:40:19 crc kubenswrapper[4980]: I1206 03:40:19.710145 4980 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-55b5d6bbdf-hgqw5" Dec 06 03:40:19 crc kubenswrapper[4980]: I1206 03:40:19.728867 4980 scope.go:117] "RemoveContainer" containerID="5193c1f55e084ebd1d67fca85e2644d601e2c7c1bf3e27ca641b00b0b502e2d8" Dec 06 03:40:19 crc kubenswrapper[4980]: I1206 03:40:19.744159 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-55b5d6bbdf-v9mg9" Dec 06 03:40:19 crc kubenswrapper[4980]: I1206 03:40:19.747472 4980 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-7565894c79-s4qkf"] Dec 06 03:40:19 crc kubenswrapper[4980]: I1206 03:40:19.756421 4980 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-7565894c79-s4qkf"] Dec 06 03:40:19 crc kubenswrapper[4980]: I1206 03:40:19.761938 4980 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-55b5d6bbdf-hgqw5"] Dec 06 03:40:19 crc kubenswrapper[4980]: I1206 03:40:19.767500 4980 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-55b5d6bbdf-hgqw5"] Dec 06 03:40:19 crc kubenswrapper[4980]: I1206 03:40:19.790134 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-677b664cd9-cbb88" Dec 06 03:40:20 crc kubenswrapper[4980]: I1206 03:40:20.042485 4980 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-55b5d6bbdf-v9mg9"] Dec 06 03:40:20 crc kubenswrapper[4980]: I1206 03:40:20.063099 4980 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-677b664cd9-cbb88"] Dec 06 03:40:20 crc kubenswrapper[4980]: I1206 03:40:20.717131 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-677b664cd9-cbb88" event={"ID":"c515dcbb-b1bb-49c4-9ea2-fc80d5fc3d9e","Type":"ContainerStarted","Data":"f8c204b02f9c44245ff1fd66e6e4353dd52b15606e7abae6baf444f9c5e3f712"} Dec 06 03:40:20 crc kubenswrapper[4980]: I1206 03:40:20.717430 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-677b664cd9-cbb88" event={"ID":"c515dcbb-b1bb-49c4-9ea2-fc80d5fc3d9e","Type":"ContainerStarted","Data":"284b9108f694d934e828d70a21b1c1dfaeb34b05559484de3d6d753eaeb1ae64"} Dec 06 03:40:20 crc kubenswrapper[4980]: I1206 03:40:20.720155 4980 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-677b664cd9-cbb88" Dec 06 03:40:20 crc kubenswrapper[4980]: I1206 03:40:20.722001 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-55b5d6bbdf-v9mg9" event={"ID":"aff033cf-643b-4060-96c2-9d65b9ceb724","Type":"ContainerStarted","Data":"530f9b9234569baeb0c0515c03307cb1f353ab52ccc8d8fe66c40e3d5fdc52b1"} Dec 06 03:40:20 crc kubenswrapper[4980]: I1206 03:40:20.722028 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-55b5d6bbdf-v9mg9" event={"ID":"aff033cf-643b-4060-96c2-9d65b9ceb724","Type":"ContainerStarted","Data":"3e7566e6233ad4fd64c8588b6638941e0a1b1d82e86a336c6fe2c6b0c447200c"} Dec 06 03:40:20 crc kubenswrapper[4980]: I1206 03:40:20.723332 4980 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-55b5d6bbdf-v9mg9" Dec 06 03:40:20 crc kubenswrapper[4980]: I1206 03:40:20.726717 4980 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-677b664cd9-cbb88" Dec 06 03:40:20 crc kubenswrapper[4980]: I1206 03:40:20.727532 4980 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-55b5d6bbdf-v9mg9" Dec 06 03:40:20 crc kubenswrapper[4980]: I1206 03:40:20.744878 4980 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-677b664cd9-cbb88" podStartSLOduration=2.744851665 podStartE2EDuration="2.744851665s" podCreationTimestamp="2025-12-06 03:40:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 03:40:20.742962507 +0000 UTC m=+379.987713778" watchObservedRunningTime="2025-12-06 03:40:20.744851665 +0000 UTC m=+379.989602936" Dec 06 03:40:20 crc kubenswrapper[4980]: I1206 03:40:20.795540 4980 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-55b5d6bbdf-v9mg9" podStartSLOduration=1.795505029 podStartE2EDuration="1.795505029s" podCreationTimestamp="2025-12-06 03:40:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 03:40:20.794014653 +0000 UTC m=+380.038765944" watchObservedRunningTime="2025-12-06 03:40:20.795505029 +0000 UTC m=+380.040256300" Dec 06 03:40:21 crc kubenswrapper[4980]: I1206 03:40:21.193630 4980 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b8cf686e-a11b-43bc-9a8d-d1063f190cde" path="/var/lib/kubelet/pods/b8cf686e-a11b-43bc-9a8d-d1063f190cde/volumes" Dec 06 03:40:21 crc kubenswrapper[4980]: I1206 03:40:21.194608 4980 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b9151d5d-626a-4872-b1f3-573394461e88" path="/var/lib/kubelet/pods/b9151d5d-626a-4872-b1f3-573394461e88/volumes" Dec 06 03:40:23 crc kubenswrapper[4980]: I1206 03:40:23.577636 4980 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-gw4xb"] Dec 06 03:40:23 crc kubenswrapper[4980]: I1206 03:40:23.578995 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-gw4xb" Dec 06 03:40:23 crc kubenswrapper[4980]: I1206 03:40:23.593409 4980 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-gw4xb"] Dec 06 03:40:23 crc kubenswrapper[4980]: I1206 03:40:23.733076 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/d5117ff8-5b5a-475b-a04e-7a18445b28cf-trusted-ca\") pod \"image-registry-66df7c8f76-gw4xb\" (UID: \"d5117ff8-5b5a-475b-a04e-7a18445b28cf\") " pod="openshift-image-registry/image-registry-66df7c8f76-gw4xb" Dec 06 03:40:23 crc kubenswrapper[4980]: I1206 03:40:23.733139 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/d5117ff8-5b5a-475b-a04e-7a18445b28cf-registry-certificates\") pod \"image-registry-66df7c8f76-gw4xb\" (UID: \"d5117ff8-5b5a-475b-a04e-7a18445b28cf\") " pod="openshift-image-registry/image-registry-66df7c8f76-gw4xb" Dec 06 03:40:23 crc kubenswrapper[4980]: I1206 03:40:23.733234 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/d5117ff8-5b5a-475b-a04e-7a18445b28cf-registry-tls\") pod \"image-registry-66df7c8f76-gw4xb\" (UID: \"d5117ff8-5b5a-475b-a04e-7a18445b28cf\") " pod="openshift-image-registry/image-registry-66df7c8f76-gw4xb" Dec 06 03:40:23 crc kubenswrapper[4980]: I1206 03:40:23.733274 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-gw4xb\" (UID: \"d5117ff8-5b5a-475b-a04e-7a18445b28cf\") " pod="openshift-image-registry/image-registry-66df7c8f76-gw4xb" Dec 06 03:40:23 crc kubenswrapper[4980]: I1206 03:40:23.733305 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/d5117ff8-5b5a-475b-a04e-7a18445b28cf-ca-trust-extracted\") pod \"image-registry-66df7c8f76-gw4xb\" (UID: \"d5117ff8-5b5a-475b-a04e-7a18445b28cf\") " pod="openshift-image-registry/image-registry-66df7c8f76-gw4xb" Dec 06 03:40:23 crc kubenswrapper[4980]: I1206 03:40:23.733327 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/d5117ff8-5b5a-475b-a04e-7a18445b28cf-bound-sa-token\") pod \"image-registry-66df7c8f76-gw4xb\" (UID: \"d5117ff8-5b5a-475b-a04e-7a18445b28cf\") " pod="openshift-image-registry/image-registry-66df7c8f76-gw4xb" Dec 06 03:40:23 crc kubenswrapper[4980]: I1206 03:40:23.733351 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-572g7\" (UniqueName: \"kubernetes.io/projected/d5117ff8-5b5a-475b-a04e-7a18445b28cf-kube-api-access-572g7\") pod \"image-registry-66df7c8f76-gw4xb\" (UID: \"d5117ff8-5b5a-475b-a04e-7a18445b28cf\") " pod="openshift-image-registry/image-registry-66df7c8f76-gw4xb" Dec 06 03:40:23 crc kubenswrapper[4980]: I1206 03:40:23.733416 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/d5117ff8-5b5a-475b-a04e-7a18445b28cf-installation-pull-secrets\") pod \"image-registry-66df7c8f76-gw4xb\" (UID: \"d5117ff8-5b5a-475b-a04e-7a18445b28cf\") " pod="openshift-image-registry/image-registry-66df7c8f76-gw4xb" Dec 06 03:40:23 crc kubenswrapper[4980]: I1206 03:40:23.759259 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-gw4xb\" (UID: \"d5117ff8-5b5a-475b-a04e-7a18445b28cf\") " pod="openshift-image-registry/image-registry-66df7c8f76-gw4xb" Dec 06 03:40:23 crc kubenswrapper[4980]: I1206 03:40:23.834920 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/d5117ff8-5b5a-475b-a04e-7a18445b28cf-installation-pull-secrets\") pod \"image-registry-66df7c8f76-gw4xb\" (UID: \"d5117ff8-5b5a-475b-a04e-7a18445b28cf\") " pod="openshift-image-registry/image-registry-66df7c8f76-gw4xb" Dec 06 03:40:23 crc kubenswrapper[4980]: I1206 03:40:23.834990 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/d5117ff8-5b5a-475b-a04e-7a18445b28cf-trusted-ca\") pod \"image-registry-66df7c8f76-gw4xb\" (UID: \"d5117ff8-5b5a-475b-a04e-7a18445b28cf\") " pod="openshift-image-registry/image-registry-66df7c8f76-gw4xb" Dec 06 03:40:23 crc kubenswrapper[4980]: I1206 03:40:23.835017 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/d5117ff8-5b5a-475b-a04e-7a18445b28cf-registry-certificates\") pod \"image-registry-66df7c8f76-gw4xb\" (UID: \"d5117ff8-5b5a-475b-a04e-7a18445b28cf\") " pod="openshift-image-registry/image-registry-66df7c8f76-gw4xb" Dec 06 03:40:23 crc kubenswrapper[4980]: I1206 03:40:23.835041 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/d5117ff8-5b5a-475b-a04e-7a18445b28cf-registry-tls\") pod \"image-registry-66df7c8f76-gw4xb\" (UID: \"d5117ff8-5b5a-475b-a04e-7a18445b28cf\") " pod="openshift-image-registry/image-registry-66df7c8f76-gw4xb" Dec 06 03:40:23 crc kubenswrapper[4980]: I1206 03:40:23.835106 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/d5117ff8-5b5a-475b-a04e-7a18445b28cf-ca-trust-extracted\") pod \"image-registry-66df7c8f76-gw4xb\" (UID: \"d5117ff8-5b5a-475b-a04e-7a18445b28cf\") " pod="openshift-image-registry/image-registry-66df7c8f76-gw4xb" Dec 06 03:40:23 crc kubenswrapper[4980]: I1206 03:40:23.835268 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/d5117ff8-5b5a-475b-a04e-7a18445b28cf-bound-sa-token\") pod \"image-registry-66df7c8f76-gw4xb\" (UID: \"d5117ff8-5b5a-475b-a04e-7a18445b28cf\") " pod="openshift-image-registry/image-registry-66df7c8f76-gw4xb" Dec 06 03:40:23 crc kubenswrapper[4980]: I1206 03:40:23.835310 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-572g7\" (UniqueName: \"kubernetes.io/projected/d5117ff8-5b5a-475b-a04e-7a18445b28cf-kube-api-access-572g7\") pod \"image-registry-66df7c8f76-gw4xb\" (UID: \"d5117ff8-5b5a-475b-a04e-7a18445b28cf\") " pod="openshift-image-registry/image-registry-66df7c8f76-gw4xb" Dec 06 03:40:23 crc kubenswrapper[4980]: I1206 03:40:23.836053 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/d5117ff8-5b5a-475b-a04e-7a18445b28cf-ca-trust-extracted\") pod \"image-registry-66df7c8f76-gw4xb\" (UID: \"d5117ff8-5b5a-475b-a04e-7a18445b28cf\") " pod="openshift-image-registry/image-registry-66df7c8f76-gw4xb" Dec 06 03:40:23 crc kubenswrapper[4980]: I1206 03:40:23.836496 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/d5117ff8-5b5a-475b-a04e-7a18445b28cf-registry-certificates\") pod \"image-registry-66df7c8f76-gw4xb\" (UID: \"d5117ff8-5b5a-475b-a04e-7a18445b28cf\") " pod="openshift-image-registry/image-registry-66df7c8f76-gw4xb" Dec 06 03:40:23 crc kubenswrapper[4980]: I1206 03:40:23.836774 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/d5117ff8-5b5a-475b-a04e-7a18445b28cf-trusted-ca\") pod \"image-registry-66df7c8f76-gw4xb\" (UID: \"d5117ff8-5b5a-475b-a04e-7a18445b28cf\") " pod="openshift-image-registry/image-registry-66df7c8f76-gw4xb" Dec 06 03:40:23 crc kubenswrapper[4980]: I1206 03:40:23.845814 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/d5117ff8-5b5a-475b-a04e-7a18445b28cf-registry-tls\") pod \"image-registry-66df7c8f76-gw4xb\" (UID: \"d5117ff8-5b5a-475b-a04e-7a18445b28cf\") " pod="openshift-image-registry/image-registry-66df7c8f76-gw4xb" Dec 06 03:40:23 crc kubenswrapper[4980]: I1206 03:40:23.845819 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/d5117ff8-5b5a-475b-a04e-7a18445b28cf-installation-pull-secrets\") pod \"image-registry-66df7c8f76-gw4xb\" (UID: \"d5117ff8-5b5a-475b-a04e-7a18445b28cf\") " pod="openshift-image-registry/image-registry-66df7c8f76-gw4xb" Dec 06 03:40:23 crc kubenswrapper[4980]: I1206 03:40:23.852586 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-572g7\" (UniqueName: \"kubernetes.io/projected/d5117ff8-5b5a-475b-a04e-7a18445b28cf-kube-api-access-572g7\") pod \"image-registry-66df7c8f76-gw4xb\" (UID: \"d5117ff8-5b5a-475b-a04e-7a18445b28cf\") " pod="openshift-image-registry/image-registry-66df7c8f76-gw4xb" Dec 06 03:40:23 crc kubenswrapper[4980]: I1206 03:40:23.855281 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/d5117ff8-5b5a-475b-a04e-7a18445b28cf-bound-sa-token\") pod \"image-registry-66df7c8f76-gw4xb\" (UID: \"d5117ff8-5b5a-475b-a04e-7a18445b28cf\") " pod="openshift-image-registry/image-registry-66df7c8f76-gw4xb" Dec 06 03:40:23 crc kubenswrapper[4980]: I1206 03:40:23.898438 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-gw4xb" Dec 06 03:40:24 crc kubenswrapper[4980]: I1206 03:40:24.165300 4980 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-gw4xb"] Dec 06 03:40:24 crc kubenswrapper[4980]: W1206 03:40:24.178996 4980 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd5117ff8_5b5a_475b_a04e_7a18445b28cf.slice/crio-77fd83aa90c4fad52b7533945a4c80dda3a22a49a6fc5826faaf3eb8761d8ca4 WatchSource:0}: Error finding container 77fd83aa90c4fad52b7533945a4c80dda3a22a49a6fc5826faaf3eb8761d8ca4: Status 404 returned error can't find the container with id 77fd83aa90c4fad52b7533945a4c80dda3a22a49a6fc5826faaf3eb8761d8ca4 Dec 06 03:40:24 crc kubenswrapper[4980]: I1206 03:40:24.757320 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-gw4xb" event={"ID":"d5117ff8-5b5a-475b-a04e-7a18445b28cf","Type":"ContainerStarted","Data":"f6f90e33c1cef079e1838dbbd5cd94974e521fadafe2c320ef74456e3a0a6def"} Dec 06 03:40:24 crc kubenswrapper[4980]: I1206 03:40:24.757856 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-gw4xb" event={"ID":"d5117ff8-5b5a-475b-a04e-7a18445b28cf","Type":"ContainerStarted","Data":"77fd83aa90c4fad52b7533945a4c80dda3a22a49a6fc5826faaf3eb8761d8ca4"} Dec 06 03:40:24 crc kubenswrapper[4980]: I1206 03:40:24.759034 4980 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-66df7c8f76-gw4xb" Dec 06 03:40:24 crc kubenswrapper[4980]: I1206 03:40:24.781300 4980 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-66df7c8f76-gw4xb" podStartSLOduration=1.781278309 podStartE2EDuration="1.781278309s" podCreationTimestamp="2025-12-06 03:40:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 03:40:24.776616405 +0000 UTC m=+384.021367666" watchObservedRunningTime="2025-12-06 03:40:24.781278309 +0000 UTC m=+384.026029580" Dec 06 03:40:25 crc kubenswrapper[4980]: I1206 03:40:25.137896 4980 patch_prober.go:28] interesting pod/machine-config-daemon-r5zfb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 03:40:25 crc kubenswrapper[4980]: I1206 03:40:25.137988 4980 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-r5zfb" podUID="320f44d1-a671-4a91-b328-a8b0fdd8f23a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 03:40:25 crc kubenswrapper[4980]: I1206 03:40:25.570479 4980 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-58d94"] Dec 06 03:40:25 crc kubenswrapper[4980]: I1206 03:40:25.571454 4980 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-58d94" podUID="e965bb7a-1bf0-4574-aae4-2a45c75fb11c" containerName="registry-server" containerID="cri-o://739e651d553aa530b056a94ecbc5131776fd3e6197e4a6840089a2c0f03e49b0" gracePeriod=30 Dec 06 03:40:25 crc kubenswrapper[4980]: I1206 03:40:25.582381 4980 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-kp68b"] Dec 06 03:40:25 crc kubenswrapper[4980]: I1206 03:40:25.582725 4980 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-kp68b" podUID="769abfb5-c18a-4cf0-a033-95b3a808867b" containerName="registry-server" containerID="cri-o://d47b4f045fefdb2c7a2a6fe135c197172906e70ebfd6c2ae3c51b3798b9982cc" gracePeriod=30 Dec 06 03:40:25 crc kubenswrapper[4980]: I1206 03:40:25.597519 4980 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-2vfzc"] Dec 06 03:40:25 crc kubenswrapper[4980]: I1206 03:40:25.597818 4980 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-2vfzc" podUID="652a7feb-30e3-4e9f-b1a9-c8ef201c658c" containerName="registry-server" containerID="cri-o://f0efc2ce0c1863787e77589a8a0afc2f9e4fdb3ce974b5bcac67705360f28d24" gracePeriod=30 Dec 06 03:40:25 crc kubenswrapper[4980]: I1206 03:40:25.612544 4980 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-mkd2m"] Dec 06 03:40:25 crc kubenswrapper[4980]: I1206 03:40:25.612911 4980 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-mkd2m" podUID="9c8b3374-77eb-44ad-b270-2263c6c23e28" containerName="registry-server" containerID="cri-o://b3b94b4e2032df3e1ef26267e391f9b2dcba67ff9306a70e1944d4df7dc01e42" gracePeriod=30 Dec 06 03:40:25 crc kubenswrapper[4980]: I1206 03:40:25.622287 4980 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-spd2v"] Dec 06 03:40:25 crc kubenswrapper[4980]: I1206 03:40:25.622748 4980 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/marketplace-operator-79b997595-spd2v" podUID="0a45bc54-be9c-4c4f-9cae-dc369d25be33" containerName="marketplace-operator" containerID="cri-o://b6c5027c91c80a97467c1e8c50318fab43b7ec468acda83d6d1f8288d5c58543" gracePeriod=30 Dec 06 03:40:25 crc kubenswrapper[4980]: I1206 03:40:25.630607 4980 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-gxcqb"] Dec 06 03:40:25 crc kubenswrapper[4980]: I1206 03:40:25.630866 4980 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-gxcqb" podUID="2ce9bc16-3d2a-4feb-82c1-0f42518196cf" containerName="registry-server" containerID="cri-o://9aa7d598bd03dc4b167375c94a61c6b2efcf6f1430a76effe832f91e012023d1" gracePeriod=30 Dec 06 03:40:25 crc kubenswrapper[4980]: I1206 03:40:25.639526 4980 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-rhhmp"] Dec 06 03:40:25 crc kubenswrapper[4980]: I1206 03:40:25.640327 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-rhhmp" Dec 06 03:40:25 crc kubenswrapper[4980]: I1206 03:40:25.651049 4980 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-xhnw8"] Dec 06 03:40:25 crc kubenswrapper[4980]: I1206 03:40:25.651341 4980 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-xhnw8" podUID="3caa51b4-a311-4498-ac9b-10c20989684a" containerName="registry-server" containerID="cri-o://e5e0adc7836a520047580410eb93beb81ac6a2dad1bc364b0c5bcf599f13cf86" gracePeriod=30 Dec 06 03:40:25 crc kubenswrapper[4980]: I1206 03:40:25.654973 4980 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-bl7hx"] Dec 06 03:40:25 crc kubenswrapper[4980]: I1206 03:40:25.655309 4980 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-bl7hx" podUID="56a102c7-f341-4076-89dc-44428e77c164" containerName="registry-server" containerID="cri-o://fdff05d7789244fc432df351c27d2f5fcb050f2030386cc3a499b53e0c252358" gracePeriod=30 Dec 06 03:40:25 crc kubenswrapper[4980]: I1206 03:40:25.670622 4980 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-rhhmp"] Dec 06 03:40:25 crc kubenswrapper[4980]: I1206 03:40:25.675675 4980 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-m9vl6"] Dec 06 03:40:25 crc kubenswrapper[4980]: I1206 03:40:25.676022 4980 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-m9vl6" podUID="43112506-9fdd-4b19-a5a5-caf864c09774" containerName="registry-server" containerID="cri-o://814e5ffce677f097050c8e056ca0ebc60c478a53dbfdeffbc4ad76fabbd8622b" gracePeriod=30 Dec 06 03:40:25 crc kubenswrapper[4980]: I1206 03:40:25.685314 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tqm24\" (UniqueName: \"kubernetes.io/projected/34c9fd26-4164-4267-8dce-7f487139e741-kube-api-access-tqm24\") pod \"marketplace-operator-79b997595-rhhmp\" (UID: \"34c9fd26-4164-4267-8dce-7f487139e741\") " pod="openshift-marketplace/marketplace-operator-79b997595-rhhmp" Dec 06 03:40:25 crc kubenswrapper[4980]: I1206 03:40:25.685366 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/34c9fd26-4164-4267-8dce-7f487139e741-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-rhhmp\" (UID: \"34c9fd26-4164-4267-8dce-7f487139e741\") " pod="openshift-marketplace/marketplace-operator-79b997595-rhhmp" Dec 06 03:40:25 crc kubenswrapper[4980]: I1206 03:40:25.685423 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/34c9fd26-4164-4267-8dce-7f487139e741-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-rhhmp\" (UID: \"34c9fd26-4164-4267-8dce-7f487139e741\") " pod="openshift-marketplace/marketplace-operator-79b997595-rhhmp" Dec 06 03:40:25 crc kubenswrapper[4980]: I1206 03:40:25.772501 4980 generic.go:334] "Generic (PLEG): container finished" podID="769abfb5-c18a-4cf0-a033-95b3a808867b" containerID="d47b4f045fefdb2c7a2a6fe135c197172906e70ebfd6c2ae3c51b3798b9982cc" exitCode=0 Dec 06 03:40:25 crc kubenswrapper[4980]: I1206 03:40:25.772600 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kp68b" event={"ID":"769abfb5-c18a-4cf0-a033-95b3a808867b","Type":"ContainerDied","Data":"d47b4f045fefdb2c7a2a6fe135c197172906e70ebfd6c2ae3c51b3798b9982cc"} Dec 06 03:40:25 crc kubenswrapper[4980]: I1206 03:40:25.775562 4980 generic.go:334] "Generic (PLEG): container finished" podID="e965bb7a-1bf0-4574-aae4-2a45c75fb11c" containerID="739e651d553aa530b056a94ecbc5131776fd3e6197e4a6840089a2c0f03e49b0" exitCode=0 Dec 06 03:40:25 crc kubenswrapper[4980]: I1206 03:40:25.775621 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-58d94" event={"ID":"e965bb7a-1bf0-4574-aae4-2a45c75fb11c","Type":"ContainerDied","Data":"739e651d553aa530b056a94ecbc5131776fd3e6197e4a6840089a2c0f03e49b0"} Dec 06 03:40:25 crc kubenswrapper[4980]: I1206 03:40:25.778846 4980 generic.go:334] "Generic (PLEG): container finished" podID="652a7feb-30e3-4e9f-b1a9-c8ef201c658c" containerID="f0efc2ce0c1863787e77589a8a0afc2f9e4fdb3ce974b5bcac67705360f28d24" exitCode=0 Dec 06 03:40:25 crc kubenswrapper[4980]: I1206 03:40:25.778989 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2vfzc" event={"ID":"652a7feb-30e3-4e9f-b1a9-c8ef201c658c","Type":"ContainerDied","Data":"f0efc2ce0c1863787e77589a8a0afc2f9e4fdb3ce974b5bcac67705360f28d24"} Dec 06 03:40:25 crc kubenswrapper[4980]: I1206 03:40:25.780120 4980 generic.go:334] "Generic (PLEG): container finished" podID="0a45bc54-be9c-4c4f-9cae-dc369d25be33" containerID="b6c5027c91c80a97467c1e8c50318fab43b7ec468acda83d6d1f8288d5c58543" exitCode=0 Dec 06 03:40:25 crc kubenswrapper[4980]: I1206 03:40:25.780170 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-spd2v" event={"ID":"0a45bc54-be9c-4c4f-9cae-dc369d25be33","Type":"ContainerDied","Data":"b6c5027c91c80a97467c1e8c50318fab43b7ec468acda83d6d1f8288d5c58543"} Dec 06 03:40:25 crc kubenswrapper[4980]: I1206 03:40:25.781648 4980 generic.go:334] "Generic (PLEG): container finished" podID="9c8b3374-77eb-44ad-b270-2263c6c23e28" containerID="b3b94b4e2032df3e1ef26267e391f9b2dcba67ff9306a70e1944d4df7dc01e42" exitCode=0 Dec 06 03:40:25 crc kubenswrapper[4980]: I1206 03:40:25.781699 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mkd2m" event={"ID":"9c8b3374-77eb-44ad-b270-2263c6c23e28","Type":"ContainerDied","Data":"b3b94b4e2032df3e1ef26267e391f9b2dcba67ff9306a70e1944d4df7dc01e42"} Dec 06 03:40:25 crc kubenswrapper[4980]: I1206 03:40:25.784481 4980 generic.go:334] "Generic (PLEG): container finished" podID="2ce9bc16-3d2a-4feb-82c1-0f42518196cf" containerID="9aa7d598bd03dc4b167375c94a61c6b2efcf6f1430a76effe832f91e012023d1" exitCode=0 Dec 06 03:40:25 crc kubenswrapper[4980]: I1206 03:40:25.785347 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-gxcqb" event={"ID":"2ce9bc16-3d2a-4feb-82c1-0f42518196cf","Type":"ContainerDied","Data":"9aa7d598bd03dc4b167375c94a61c6b2efcf6f1430a76effe832f91e012023d1"} Dec 06 03:40:25 crc kubenswrapper[4980]: I1206 03:40:25.801949 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tqm24\" (UniqueName: \"kubernetes.io/projected/34c9fd26-4164-4267-8dce-7f487139e741-kube-api-access-tqm24\") pod \"marketplace-operator-79b997595-rhhmp\" (UID: \"34c9fd26-4164-4267-8dce-7f487139e741\") " pod="openshift-marketplace/marketplace-operator-79b997595-rhhmp" Dec 06 03:40:25 crc kubenswrapper[4980]: I1206 03:40:25.801993 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/34c9fd26-4164-4267-8dce-7f487139e741-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-rhhmp\" (UID: \"34c9fd26-4164-4267-8dce-7f487139e741\") " pod="openshift-marketplace/marketplace-operator-79b997595-rhhmp" Dec 06 03:40:25 crc kubenswrapper[4980]: I1206 03:40:25.802014 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/34c9fd26-4164-4267-8dce-7f487139e741-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-rhhmp\" (UID: \"34c9fd26-4164-4267-8dce-7f487139e741\") " pod="openshift-marketplace/marketplace-operator-79b997595-rhhmp" Dec 06 03:40:25 crc kubenswrapper[4980]: I1206 03:40:25.806841 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/34c9fd26-4164-4267-8dce-7f487139e741-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-rhhmp\" (UID: \"34c9fd26-4164-4267-8dce-7f487139e741\") " pod="openshift-marketplace/marketplace-operator-79b997595-rhhmp" Dec 06 03:40:25 crc kubenswrapper[4980]: I1206 03:40:25.812957 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/34c9fd26-4164-4267-8dce-7f487139e741-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-rhhmp\" (UID: \"34c9fd26-4164-4267-8dce-7f487139e741\") " pod="openshift-marketplace/marketplace-operator-79b997595-rhhmp" Dec 06 03:40:25 crc kubenswrapper[4980]: I1206 03:40:25.820622 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tqm24\" (UniqueName: \"kubernetes.io/projected/34c9fd26-4164-4267-8dce-7f487139e741-kube-api-access-tqm24\") pod \"marketplace-operator-79b997595-rhhmp\" (UID: \"34c9fd26-4164-4267-8dce-7f487139e741\") " pod="openshift-marketplace/marketplace-operator-79b997595-rhhmp" Dec 06 03:40:26 crc kubenswrapper[4980]: I1206 03:40:26.401892 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-rhhmp" Dec 06 03:40:26 crc kubenswrapper[4980]: I1206 03:40:26.415414 4980 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-58d94" Dec 06 03:40:26 crc kubenswrapper[4980]: I1206 03:40:26.510681 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e965bb7a-1bf0-4574-aae4-2a45c75fb11c-utilities\") pod \"e965bb7a-1bf0-4574-aae4-2a45c75fb11c\" (UID: \"e965bb7a-1bf0-4574-aae4-2a45c75fb11c\") " Dec 06 03:40:26 crc kubenswrapper[4980]: I1206 03:40:26.510780 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4qvrf\" (UniqueName: \"kubernetes.io/projected/e965bb7a-1bf0-4574-aae4-2a45c75fb11c-kube-api-access-4qvrf\") pod \"e965bb7a-1bf0-4574-aae4-2a45c75fb11c\" (UID: \"e965bb7a-1bf0-4574-aae4-2a45c75fb11c\") " Dec 06 03:40:26 crc kubenswrapper[4980]: I1206 03:40:26.510840 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e965bb7a-1bf0-4574-aae4-2a45c75fb11c-catalog-content\") pod \"e965bb7a-1bf0-4574-aae4-2a45c75fb11c\" (UID: \"e965bb7a-1bf0-4574-aae4-2a45c75fb11c\") " Dec 06 03:40:26 crc kubenswrapper[4980]: I1206 03:40:26.514767 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e965bb7a-1bf0-4574-aae4-2a45c75fb11c-utilities" (OuterVolumeSpecName: "utilities") pod "e965bb7a-1bf0-4574-aae4-2a45c75fb11c" (UID: "e965bb7a-1bf0-4574-aae4-2a45c75fb11c"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 03:40:26 crc kubenswrapper[4980]: I1206 03:40:26.525937 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e965bb7a-1bf0-4574-aae4-2a45c75fb11c-kube-api-access-4qvrf" (OuterVolumeSpecName: "kube-api-access-4qvrf") pod "e965bb7a-1bf0-4574-aae4-2a45c75fb11c" (UID: "e965bb7a-1bf0-4574-aae4-2a45c75fb11c"). InnerVolumeSpecName "kube-api-access-4qvrf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 03:40:26 crc kubenswrapper[4980]: I1206 03:40:26.539908 4980 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-gxcqb" Dec 06 03:40:26 crc kubenswrapper[4980]: I1206 03:40:26.541204 4980 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-2vfzc" Dec 06 03:40:26 crc kubenswrapper[4980]: I1206 03:40:26.542873 4980 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-mkd2m" Dec 06 03:40:26 crc kubenswrapper[4980]: I1206 03:40:26.547222 4980 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-bl7hx" Dec 06 03:40:26 crc kubenswrapper[4980]: I1206 03:40:26.554292 4980 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-spd2v" Dec 06 03:40:26 crc kubenswrapper[4980]: I1206 03:40:26.559858 4980 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-kp68b" Dec 06 03:40:26 crc kubenswrapper[4980]: I1206 03:40:26.564239 4980 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-m9vl6" Dec 06 03:40:26 crc kubenswrapper[4980]: I1206 03:40:26.573663 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e965bb7a-1bf0-4574-aae4-2a45c75fb11c-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "e965bb7a-1bf0-4574-aae4-2a45c75fb11c" (UID: "e965bb7a-1bf0-4574-aae4-2a45c75fb11c"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 03:40:26 crc kubenswrapper[4980]: I1206 03:40:26.580967 4980 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-xhnw8" Dec 06 03:40:26 crc kubenswrapper[4980]: I1206 03:40:26.612648 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3caa51b4-a311-4498-ac9b-10c20989684a-utilities\") pod \"3caa51b4-a311-4498-ac9b-10c20989684a\" (UID: \"3caa51b4-a311-4498-ac9b-10c20989684a\") " Dec 06 03:40:26 crc kubenswrapper[4980]: I1206 03:40:26.612689 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9c8b3374-77eb-44ad-b270-2263c6c23e28-utilities\") pod \"9c8b3374-77eb-44ad-b270-2263c6c23e28\" (UID: \"9c8b3374-77eb-44ad-b270-2263c6c23e28\") " Dec 06 03:40:26 crc kubenswrapper[4980]: I1206 03:40:26.612706 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/769abfb5-c18a-4cf0-a033-95b3a808867b-catalog-content\") pod \"769abfb5-c18a-4cf0-a033-95b3a808867b\" (UID: \"769abfb5-c18a-4cf0-a033-95b3a808867b\") " Dec 06 03:40:26 crc kubenswrapper[4980]: I1206 03:40:26.612731 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9c8b3374-77eb-44ad-b270-2263c6c23e28-catalog-content\") pod \"9c8b3374-77eb-44ad-b270-2263c6c23e28\" (UID: \"9c8b3374-77eb-44ad-b270-2263c6c23e28\") " Dec 06 03:40:26 crc kubenswrapper[4980]: I1206 03:40:26.612760 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/43112506-9fdd-4b19-a5a5-caf864c09774-catalog-content\") pod \"43112506-9fdd-4b19-a5a5-caf864c09774\" (UID: \"43112506-9fdd-4b19-a5a5-caf864c09774\") " Dec 06 03:40:26 crc kubenswrapper[4980]: I1206 03:40:26.612781 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cnnh8\" (UniqueName: \"kubernetes.io/projected/769abfb5-c18a-4cf0-a033-95b3a808867b-kube-api-access-cnnh8\") pod \"769abfb5-c18a-4cf0-a033-95b3a808867b\" (UID: \"769abfb5-c18a-4cf0-a033-95b3a808867b\") " Dec 06 03:40:26 crc kubenswrapper[4980]: I1206 03:40:26.612797 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/56a102c7-f341-4076-89dc-44428e77c164-catalog-content\") pod \"56a102c7-f341-4076-89dc-44428e77c164\" (UID: \"56a102c7-f341-4076-89dc-44428e77c164\") " Dec 06 03:40:26 crc kubenswrapper[4980]: I1206 03:40:26.612831 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2ce9bc16-3d2a-4feb-82c1-0f42518196cf-utilities\") pod \"2ce9bc16-3d2a-4feb-82c1-0f42518196cf\" (UID: \"2ce9bc16-3d2a-4feb-82c1-0f42518196cf\") " Dec 06 03:40:26 crc kubenswrapper[4980]: I1206 03:40:26.612856 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2ce9bc16-3d2a-4feb-82c1-0f42518196cf-catalog-content\") pod \"2ce9bc16-3d2a-4feb-82c1-0f42518196cf\" (UID: \"2ce9bc16-3d2a-4feb-82c1-0f42518196cf\") " Dec 06 03:40:26 crc kubenswrapper[4980]: I1206 03:40:26.612885 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b9kxc\" (UniqueName: \"kubernetes.io/projected/2ce9bc16-3d2a-4feb-82c1-0f42518196cf-kube-api-access-b9kxc\") pod \"2ce9bc16-3d2a-4feb-82c1-0f42518196cf\" (UID: \"2ce9bc16-3d2a-4feb-82c1-0f42518196cf\") " Dec 06 03:40:26 crc kubenswrapper[4980]: I1206 03:40:26.612913 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vpb28\" (UniqueName: \"kubernetes.io/projected/56a102c7-f341-4076-89dc-44428e77c164-kube-api-access-vpb28\") pod \"56a102c7-f341-4076-89dc-44428e77c164\" (UID: \"56a102c7-f341-4076-89dc-44428e77c164\") " Dec 06 03:40:26 crc kubenswrapper[4980]: I1206 03:40:26.612933 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hwq2v\" (UniqueName: \"kubernetes.io/projected/43112506-9fdd-4b19-a5a5-caf864c09774-kube-api-access-hwq2v\") pod \"43112506-9fdd-4b19-a5a5-caf864c09774\" (UID: \"43112506-9fdd-4b19-a5a5-caf864c09774\") " Dec 06 03:40:26 crc kubenswrapper[4980]: I1206 03:40:26.612954 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/0a45bc54-be9c-4c4f-9cae-dc369d25be33-marketplace-trusted-ca\") pod \"0a45bc54-be9c-4c4f-9cae-dc369d25be33\" (UID: \"0a45bc54-be9c-4c4f-9cae-dc369d25be33\") " Dec 06 03:40:26 crc kubenswrapper[4980]: I1206 03:40:26.612980 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zgc7q\" (UniqueName: \"kubernetes.io/projected/3caa51b4-a311-4498-ac9b-10c20989684a-kube-api-access-zgc7q\") pod \"3caa51b4-a311-4498-ac9b-10c20989684a\" (UID: \"3caa51b4-a311-4498-ac9b-10c20989684a\") " Dec 06 03:40:26 crc kubenswrapper[4980]: I1206 03:40:26.613001 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3caa51b4-a311-4498-ac9b-10c20989684a-catalog-content\") pod \"3caa51b4-a311-4498-ac9b-10c20989684a\" (UID: \"3caa51b4-a311-4498-ac9b-10c20989684a\") " Dec 06 03:40:26 crc kubenswrapper[4980]: I1206 03:40:26.613033 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ppcll\" (UniqueName: \"kubernetes.io/projected/9c8b3374-77eb-44ad-b270-2263c6c23e28-kube-api-access-ppcll\") pod \"9c8b3374-77eb-44ad-b270-2263c6c23e28\" (UID: \"9c8b3374-77eb-44ad-b270-2263c6c23e28\") " Dec 06 03:40:26 crc kubenswrapper[4980]: I1206 03:40:26.613048 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/769abfb5-c18a-4cf0-a033-95b3a808867b-utilities\") pod \"769abfb5-c18a-4cf0-a033-95b3a808867b\" (UID: \"769abfb5-c18a-4cf0-a033-95b3a808867b\") " Dec 06 03:40:26 crc kubenswrapper[4980]: I1206 03:40:26.613065 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/652a7feb-30e3-4e9f-b1a9-c8ef201c658c-catalog-content\") pod \"652a7feb-30e3-4e9f-b1a9-c8ef201c658c\" (UID: \"652a7feb-30e3-4e9f-b1a9-c8ef201c658c\") " Dec 06 03:40:26 crc kubenswrapper[4980]: I1206 03:40:26.613096 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/43112506-9fdd-4b19-a5a5-caf864c09774-utilities\") pod \"43112506-9fdd-4b19-a5a5-caf864c09774\" (UID: \"43112506-9fdd-4b19-a5a5-caf864c09774\") " Dec 06 03:40:26 crc kubenswrapper[4980]: I1206 03:40:26.613110 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/56a102c7-f341-4076-89dc-44428e77c164-utilities\") pod \"56a102c7-f341-4076-89dc-44428e77c164\" (UID: \"56a102c7-f341-4076-89dc-44428e77c164\") " Dec 06 03:40:26 crc kubenswrapper[4980]: I1206 03:40:26.613126 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wxr7g\" (UniqueName: \"kubernetes.io/projected/652a7feb-30e3-4e9f-b1a9-c8ef201c658c-kube-api-access-wxr7g\") pod \"652a7feb-30e3-4e9f-b1a9-c8ef201c658c\" (UID: \"652a7feb-30e3-4e9f-b1a9-c8ef201c658c\") " Dec 06 03:40:26 crc kubenswrapper[4980]: I1206 03:40:26.613145 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/652a7feb-30e3-4e9f-b1a9-c8ef201c658c-utilities\") pod \"652a7feb-30e3-4e9f-b1a9-c8ef201c658c\" (UID: \"652a7feb-30e3-4e9f-b1a9-c8ef201c658c\") " Dec 06 03:40:26 crc kubenswrapper[4980]: I1206 03:40:26.613165 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/0a45bc54-be9c-4c4f-9cae-dc369d25be33-marketplace-operator-metrics\") pod \"0a45bc54-be9c-4c4f-9cae-dc369d25be33\" (UID: \"0a45bc54-be9c-4c4f-9cae-dc369d25be33\") " Dec 06 03:40:26 crc kubenswrapper[4980]: I1206 03:40:26.613193 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f5l5z\" (UniqueName: \"kubernetes.io/projected/0a45bc54-be9c-4c4f-9cae-dc369d25be33-kube-api-access-f5l5z\") pod \"0a45bc54-be9c-4c4f-9cae-dc369d25be33\" (UID: \"0a45bc54-be9c-4c4f-9cae-dc369d25be33\") " Dec 06 03:40:26 crc kubenswrapper[4980]: I1206 03:40:26.613499 4980 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e965bb7a-1bf0-4574-aae4-2a45c75fb11c-utilities\") on node \"crc\" DevicePath \"\"" Dec 06 03:40:26 crc kubenswrapper[4980]: I1206 03:40:26.613528 4980 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4qvrf\" (UniqueName: \"kubernetes.io/projected/e965bb7a-1bf0-4574-aae4-2a45c75fb11c-kube-api-access-4qvrf\") on node \"crc\" DevicePath \"\"" Dec 06 03:40:26 crc kubenswrapper[4980]: I1206 03:40:26.613538 4980 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e965bb7a-1bf0-4574-aae4-2a45c75fb11c-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 06 03:40:26 crc kubenswrapper[4980]: I1206 03:40:26.613564 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3caa51b4-a311-4498-ac9b-10c20989684a-utilities" (OuterVolumeSpecName: "utilities") pod "3caa51b4-a311-4498-ac9b-10c20989684a" (UID: "3caa51b4-a311-4498-ac9b-10c20989684a"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 03:40:26 crc kubenswrapper[4980]: I1206 03:40:26.616416 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0a45bc54-be9c-4c4f-9cae-dc369d25be33-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "0a45bc54-be9c-4c4f-9cae-dc369d25be33" (UID: "0a45bc54-be9c-4c4f-9cae-dc369d25be33"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 03:40:26 crc kubenswrapper[4980]: I1206 03:40:26.617261 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9c8b3374-77eb-44ad-b270-2263c6c23e28-utilities" (OuterVolumeSpecName: "utilities") pod "9c8b3374-77eb-44ad-b270-2263c6c23e28" (UID: "9c8b3374-77eb-44ad-b270-2263c6c23e28"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 03:40:26 crc kubenswrapper[4980]: I1206 03:40:26.618042 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/43112506-9fdd-4b19-a5a5-caf864c09774-utilities" (OuterVolumeSpecName: "utilities") pod "43112506-9fdd-4b19-a5a5-caf864c09774" (UID: "43112506-9fdd-4b19-a5a5-caf864c09774"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 03:40:26 crc kubenswrapper[4980]: I1206 03:40:26.618087 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/652a7feb-30e3-4e9f-b1a9-c8ef201c658c-utilities" (OuterVolumeSpecName: "utilities") pod "652a7feb-30e3-4e9f-b1a9-c8ef201c658c" (UID: "652a7feb-30e3-4e9f-b1a9-c8ef201c658c"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 03:40:26 crc kubenswrapper[4980]: I1206 03:40:26.619120 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2ce9bc16-3d2a-4feb-82c1-0f42518196cf-utilities" (OuterVolumeSpecName: "utilities") pod "2ce9bc16-3d2a-4feb-82c1-0f42518196cf" (UID: "2ce9bc16-3d2a-4feb-82c1-0f42518196cf"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 03:40:26 crc kubenswrapper[4980]: I1206 03:40:26.630883 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/769abfb5-c18a-4cf0-a033-95b3a808867b-utilities" (OuterVolumeSpecName: "utilities") pod "769abfb5-c18a-4cf0-a033-95b3a808867b" (UID: "769abfb5-c18a-4cf0-a033-95b3a808867b"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 03:40:26 crc kubenswrapper[4980]: I1206 03:40:26.631785 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/56a102c7-f341-4076-89dc-44428e77c164-utilities" (OuterVolumeSpecName: "utilities") pod "56a102c7-f341-4076-89dc-44428e77c164" (UID: "56a102c7-f341-4076-89dc-44428e77c164"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 03:40:26 crc kubenswrapper[4980]: I1206 03:40:26.636128 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9c8b3374-77eb-44ad-b270-2263c6c23e28-kube-api-access-ppcll" (OuterVolumeSpecName: "kube-api-access-ppcll") pod "9c8b3374-77eb-44ad-b270-2263c6c23e28" (UID: "9c8b3374-77eb-44ad-b270-2263c6c23e28"). InnerVolumeSpecName "kube-api-access-ppcll". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 03:40:26 crc kubenswrapper[4980]: I1206 03:40:26.636237 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0a45bc54-be9c-4c4f-9cae-dc369d25be33-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "0a45bc54-be9c-4c4f-9cae-dc369d25be33" (UID: "0a45bc54-be9c-4c4f-9cae-dc369d25be33"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 03:40:26 crc kubenswrapper[4980]: I1206 03:40:26.637294 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/56a102c7-f341-4076-89dc-44428e77c164-kube-api-access-vpb28" (OuterVolumeSpecName: "kube-api-access-vpb28") pod "56a102c7-f341-4076-89dc-44428e77c164" (UID: "56a102c7-f341-4076-89dc-44428e77c164"). InnerVolumeSpecName "kube-api-access-vpb28". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 03:40:26 crc kubenswrapper[4980]: I1206 03:40:26.638084 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3caa51b4-a311-4498-ac9b-10c20989684a-kube-api-access-zgc7q" (OuterVolumeSpecName: "kube-api-access-zgc7q") pod "3caa51b4-a311-4498-ac9b-10c20989684a" (UID: "3caa51b4-a311-4498-ac9b-10c20989684a"). InnerVolumeSpecName "kube-api-access-zgc7q". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 03:40:26 crc kubenswrapper[4980]: I1206 03:40:26.638375 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/769abfb5-c18a-4cf0-a033-95b3a808867b-kube-api-access-cnnh8" (OuterVolumeSpecName: "kube-api-access-cnnh8") pod "769abfb5-c18a-4cf0-a033-95b3a808867b" (UID: "769abfb5-c18a-4cf0-a033-95b3a808867b"). InnerVolumeSpecName "kube-api-access-cnnh8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 03:40:26 crc kubenswrapper[4980]: I1206 03:40:26.640952 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0a45bc54-be9c-4c4f-9cae-dc369d25be33-kube-api-access-f5l5z" (OuterVolumeSpecName: "kube-api-access-f5l5z") pod "0a45bc54-be9c-4c4f-9cae-dc369d25be33" (UID: "0a45bc54-be9c-4c4f-9cae-dc369d25be33"). InnerVolumeSpecName "kube-api-access-f5l5z". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 03:40:26 crc kubenswrapper[4980]: I1206 03:40:26.644846 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2ce9bc16-3d2a-4feb-82c1-0f42518196cf-kube-api-access-b9kxc" (OuterVolumeSpecName: "kube-api-access-b9kxc") pod "2ce9bc16-3d2a-4feb-82c1-0f42518196cf" (UID: "2ce9bc16-3d2a-4feb-82c1-0f42518196cf"). InnerVolumeSpecName "kube-api-access-b9kxc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 03:40:26 crc kubenswrapper[4980]: I1206 03:40:26.650713 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/43112506-9fdd-4b19-a5a5-caf864c09774-kube-api-access-hwq2v" (OuterVolumeSpecName: "kube-api-access-hwq2v") pod "43112506-9fdd-4b19-a5a5-caf864c09774" (UID: "43112506-9fdd-4b19-a5a5-caf864c09774"). InnerVolumeSpecName "kube-api-access-hwq2v". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 03:40:26 crc kubenswrapper[4980]: I1206 03:40:26.661234 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/652a7feb-30e3-4e9f-b1a9-c8ef201c658c-kube-api-access-wxr7g" (OuterVolumeSpecName: "kube-api-access-wxr7g") pod "652a7feb-30e3-4e9f-b1a9-c8ef201c658c" (UID: "652a7feb-30e3-4e9f-b1a9-c8ef201c658c"). InnerVolumeSpecName "kube-api-access-wxr7g". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 03:40:26 crc kubenswrapper[4980]: I1206 03:40:26.663011 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2ce9bc16-3d2a-4feb-82c1-0f42518196cf-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "2ce9bc16-3d2a-4feb-82c1-0f42518196cf" (UID: "2ce9bc16-3d2a-4feb-82c1-0f42518196cf"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 03:40:26 crc kubenswrapper[4980]: I1206 03:40:26.688645 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3caa51b4-a311-4498-ac9b-10c20989684a-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "3caa51b4-a311-4498-ac9b-10c20989684a" (UID: "3caa51b4-a311-4498-ac9b-10c20989684a"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 03:40:26 crc kubenswrapper[4980]: I1206 03:40:26.715611 4980 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2ce9bc16-3d2a-4feb-82c1-0f42518196cf-utilities\") on node \"crc\" DevicePath \"\"" Dec 06 03:40:26 crc kubenswrapper[4980]: I1206 03:40:26.715652 4980 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b9kxc\" (UniqueName: \"kubernetes.io/projected/2ce9bc16-3d2a-4feb-82c1-0f42518196cf-kube-api-access-b9kxc\") on node \"crc\" DevicePath \"\"" Dec 06 03:40:26 crc kubenswrapper[4980]: I1206 03:40:26.715665 4980 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2ce9bc16-3d2a-4feb-82c1-0f42518196cf-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 06 03:40:26 crc kubenswrapper[4980]: I1206 03:40:26.715679 4980 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vpb28\" (UniqueName: \"kubernetes.io/projected/56a102c7-f341-4076-89dc-44428e77c164-kube-api-access-vpb28\") on node \"crc\" DevicePath \"\"" Dec 06 03:40:26 crc kubenswrapper[4980]: I1206 03:40:26.715691 4980 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hwq2v\" (UniqueName: \"kubernetes.io/projected/43112506-9fdd-4b19-a5a5-caf864c09774-kube-api-access-hwq2v\") on node \"crc\" DevicePath \"\"" Dec 06 03:40:26 crc kubenswrapper[4980]: I1206 03:40:26.715702 4980 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/0a45bc54-be9c-4c4f-9cae-dc369d25be33-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 06 03:40:26 crc kubenswrapper[4980]: I1206 03:40:26.715714 4980 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zgc7q\" (UniqueName: \"kubernetes.io/projected/3caa51b4-a311-4498-ac9b-10c20989684a-kube-api-access-zgc7q\") on node \"crc\" DevicePath \"\"" Dec 06 03:40:26 crc kubenswrapper[4980]: I1206 03:40:26.715727 4980 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3caa51b4-a311-4498-ac9b-10c20989684a-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 06 03:40:26 crc kubenswrapper[4980]: I1206 03:40:26.715737 4980 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ppcll\" (UniqueName: \"kubernetes.io/projected/9c8b3374-77eb-44ad-b270-2263c6c23e28-kube-api-access-ppcll\") on node \"crc\" DevicePath \"\"" Dec 06 03:40:26 crc kubenswrapper[4980]: I1206 03:40:26.715747 4980 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/769abfb5-c18a-4cf0-a033-95b3a808867b-utilities\") on node \"crc\" DevicePath \"\"" Dec 06 03:40:26 crc kubenswrapper[4980]: I1206 03:40:26.715758 4980 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/43112506-9fdd-4b19-a5a5-caf864c09774-utilities\") on node \"crc\" DevicePath \"\"" Dec 06 03:40:26 crc kubenswrapper[4980]: I1206 03:40:26.715768 4980 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/56a102c7-f341-4076-89dc-44428e77c164-utilities\") on node \"crc\" DevicePath \"\"" Dec 06 03:40:26 crc kubenswrapper[4980]: I1206 03:40:26.715779 4980 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wxr7g\" (UniqueName: \"kubernetes.io/projected/652a7feb-30e3-4e9f-b1a9-c8ef201c658c-kube-api-access-wxr7g\") on node \"crc\" DevicePath \"\"" Dec 06 03:40:26 crc kubenswrapper[4980]: I1206 03:40:26.715789 4980 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/652a7feb-30e3-4e9f-b1a9-c8ef201c658c-utilities\") on node \"crc\" DevicePath \"\"" Dec 06 03:40:26 crc kubenswrapper[4980]: I1206 03:40:26.715801 4980 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/0a45bc54-be9c-4c4f-9cae-dc369d25be33-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Dec 06 03:40:26 crc kubenswrapper[4980]: I1206 03:40:26.715813 4980 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f5l5z\" (UniqueName: \"kubernetes.io/projected/0a45bc54-be9c-4c4f-9cae-dc369d25be33-kube-api-access-f5l5z\") on node \"crc\" DevicePath \"\"" Dec 06 03:40:26 crc kubenswrapper[4980]: I1206 03:40:26.715823 4980 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3caa51b4-a311-4498-ac9b-10c20989684a-utilities\") on node \"crc\" DevicePath \"\"" Dec 06 03:40:26 crc kubenswrapper[4980]: I1206 03:40:26.715833 4980 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9c8b3374-77eb-44ad-b270-2263c6c23e28-utilities\") on node \"crc\" DevicePath \"\"" Dec 06 03:40:26 crc kubenswrapper[4980]: I1206 03:40:26.715844 4980 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cnnh8\" (UniqueName: \"kubernetes.io/projected/769abfb5-c18a-4cf0-a033-95b3a808867b-kube-api-access-cnnh8\") on node \"crc\" DevicePath \"\"" Dec 06 03:40:26 crc kubenswrapper[4980]: I1206 03:40:26.740270 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/769abfb5-c18a-4cf0-a033-95b3a808867b-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "769abfb5-c18a-4cf0-a033-95b3a808867b" (UID: "769abfb5-c18a-4cf0-a033-95b3a808867b"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 03:40:26 crc kubenswrapper[4980]: I1206 03:40:26.752737 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9c8b3374-77eb-44ad-b270-2263c6c23e28-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "9c8b3374-77eb-44ad-b270-2263c6c23e28" (UID: "9c8b3374-77eb-44ad-b270-2263c6c23e28"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 03:40:26 crc kubenswrapper[4980]: I1206 03:40:26.780426 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/43112506-9fdd-4b19-a5a5-caf864c09774-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "43112506-9fdd-4b19-a5a5-caf864c09774" (UID: "43112506-9fdd-4b19-a5a5-caf864c09774"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 03:40:26 crc kubenswrapper[4980]: I1206 03:40:26.798850 4980 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-mkd2m" Dec 06 03:40:26 crc kubenswrapper[4980]: I1206 03:40:26.799031 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/652a7feb-30e3-4e9f-b1a9-c8ef201c658c-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "652a7feb-30e3-4e9f-b1a9-c8ef201c658c" (UID: "652a7feb-30e3-4e9f-b1a9-c8ef201c658c"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 03:40:26 crc kubenswrapper[4980]: I1206 03:40:26.799581 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mkd2m" event={"ID":"9c8b3374-77eb-44ad-b270-2263c6c23e28","Type":"ContainerDied","Data":"48c1a73f4a7ea2d6254b7b0f4f0d5c79560a3376f75717817c4ff9dc9e88b8cd"} Dec 06 03:40:26 crc kubenswrapper[4980]: I1206 03:40:26.799645 4980 scope.go:117] "RemoveContainer" containerID="b3b94b4e2032df3e1ef26267e391f9b2dcba67ff9306a70e1944d4df7dc01e42" Dec 06 03:40:26 crc kubenswrapper[4980]: I1206 03:40:26.814375 4980 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-kp68b" Dec 06 03:40:26 crc kubenswrapper[4980]: I1206 03:40:26.814999 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kp68b" event={"ID":"769abfb5-c18a-4cf0-a033-95b3a808867b","Type":"ContainerDied","Data":"87b491bed7e9e4c176b0bb2f09b26aa5929116e0c3dd3a76189f34e09567bc1a"} Dec 06 03:40:26 crc kubenswrapper[4980]: I1206 03:40:26.817041 4980 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/652a7feb-30e3-4e9f-b1a9-c8ef201c658c-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 06 03:40:26 crc kubenswrapper[4980]: I1206 03:40:26.817214 4980 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/769abfb5-c18a-4cf0-a033-95b3a808867b-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 06 03:40:26 crc kubenswrapper[4980]: I1206 03:40:26.817226 4980 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9c8b3374-77eb-44ad-b270-2263c6c23e28-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 06 03:40:26 crc kubenswrapper[4980]: I1206 03:40:26.817234 4980 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/43112506-9fdd-4b19-a5a5-caf864c09774-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 06 03:40:26 crc kubenswrapper[4980]: I1206 03:40:26.818644 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-58d94" event={"ID":"e965bb7a-1bf0-4574-aae4-2a45c75fb11c","Type":"ContainerDied","Data":"5c48468877394b6180088d640646f461561cffea4774804c561aae3e0fda3e81"} Dec 06 03:40:26 crc kubenswrapper[4980]: I1206 03:40:26.818978 4980 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-58d94" Dec 06 03:40:26 crc kubenswrapper[4980]: I1206 03:40:26.821236 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-spd2v" event={"ID":"0a45bc54-be9c-4c4f-9cae-dc369d25be33","Type":"ContainerDied","Data":"7a8808a8c7f810b0ac36ffaffe6515bcff3c06970f3d138268138750f476b32d"} Dec 06 03:40:26 crc kubenswrapper[4980]: I1206 03:40:26.821316 4980 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-spd2v" Dec 06 03:40:26 crc kubenswrapper[4980]: I1206 03:40:26.823657 4980 scope.go:117] "RemoveContainer" containerID="402a4b6521a47e3cfb5a8ca7a94ee76b83c72ac0e2c6645c91e7fcd6bee0b4bc" Dec 06 03:40:26 crc kubenswrapper[4980]: I1206 03:40:26.826012 4980 generic.go:334] "Generic (PLEG): container finished" podID="56a102c7-f341-4076-89dc-44428e77c164" containerID="fdff05d7789244fc432df351c27d2f5fcb050f2030386cc3a499b53e0c252358" exitCode=0 Dec 06 03:40:26 crc kubenswrapper[4980]: I1206 03:40:26.826106 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bl7hx" event={"ID":"56a102c7-f341-4076-89dc-44428e77c164","Type":"ContainerDied","Data":"fdff05d7789244fc432df351c27d2f5fcb050f2030386cc3a499b53e0c252358"} Dec 06 03:40:26 crc kubenswrapper[4980]: I1206 03:40:26.826158 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bl7hx" event={"ID":"56a102c7-f341-4076-89dc-44428e77c164","Type":"ContainerDied","Data":"ed86090995deeef27f17707d16552c129a494499b8a80c3ec72ae12c80cc31b8"} Dec 06 03:40:26 crc kubenswrapper[4980]: I1206 03:40:26.826267 4980 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-bl7hx" Dec 06 03:40:26 crc kubenswrapper[4980]: I1206 03:40:26.852140 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-gxcqb" event={"ID":"2ce9bc16-3d2a-4feb-82c1-0f42518196cf","Type":"ContainerDied","Data":"88edfcbb22a3aa98874380591fdb3d5827b4fb84a7c528e82e509639ce7804f8"} Dec 06 03:40:26 crc kubenswrapper[4980]: I1206 03:40:26.852802 4980 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-gxcqb" Dec 06 03:40:26 crc kubenswrapper[4980]: I1206 03:40:26.856001 4980 generic.go:334] "Generic (PLEG): container finished" podID="43112506-9fdd-4b19-a5a5-caf864c09774" containerID="814e5ffce677f097050c8e056ca0ebc60c478a53dbfdeffbc4ad76fabbd8622b" exitCode=0 Dec 06 03:40:26 crc kubenswrapper[4980]: I1206 03:40:26.856111 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-m9vl6" event={"ID":"43112506-9fdd-4b19-a5a5-caf864c09774","Type":"ContainerDied","Data":"814e5ffce677f097050c8e056ca0ebc60c478a53dbfdeffbc4ad76fabbd8622b"} Dec 06 03:40:26 crc kubenswrapper[4980]: I1206 03:40:26.856144 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-m9vl6" event={"ID":"43112506-9fdd-4b19-a5a5-caf864c09774","Type":"ContainerDied","Data":"29df7b7fb5fdf33ae3bf5ebbe15b646210dafb41c0c00f8e8ab3338bc1294d88"} Dec 06 03:40:26 crc kubenswrapper[4980]: I1206 03:40:26.856360 4980 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-m9vl6" Dec 06 03:40:26 crc kubenswrapper[4980]: I1206 03:40:26.857778 4980 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-mkd2m"] Dec 06 03:40:26 crc kubenswrapper[4980]: I1206 03:40:26.870793 4980 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-mkd2m"] Dec 06 03:40:26 crc kubenswrapper[4980]: I1206 03:40:26.908004 4980 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-2vfzc" Dec 06 03:40:26 crc kubenswrapper[4980]: I1206 03:40:26.908002 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2vfzc" event={"ID":"652a7feb-30e3-4e9f-b1a9-c8ef201c658c","Type":"ContainerDied","Data":"4cf3f79b3da4c6177b583e0ef3da23b5cd48f26fa556daa5657f94db53da4a22"} Dec 06 03:40:26 crc kubenswrapper[4980]: I1206 03:40:26.917088 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/56a102c7-f341-4076-89dc-44428e77c164-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "56a102c7-f341-4076-89dc-44428e77c164" (UID: "56a102c7-f341-4076-89dc-44428e77c164"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 03:40:26 crc kubenswrapper[4980]: I1206 03:40:26.918496 4980 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-rhhmp"] Dec 06 03:40:26 crc kubenswrapper[4980]: I1206 03:40:26.920412 4980 generic.go:334] "Generic (PLEG): container finished" podID="3caa51b4-a311-4498-ac9b-10c20989684a" containerID="e5e0adc7836a520047580410eb93beb81ac6a2dad1bc364b0c5bcf599f13cf86" exitCode=0 Dec 06 03:40:26 crc kubenswrapper[4980]: I1206 03:40:26.920482 4980 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-xhnw8" Dec 06 03:40:26 crc kubenswrapper[4980]: I1206 03:40:26.920547 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xhnw8" event={"ID":"3caa51b4-a311-4498-ac9b-10c20989684a","Type":"ContainerDied","Data":"e5e0adc7836a520047580410eb93beb81ac6a2dad1bc364b0c5bcf599f13cf86"} Dec 06 03:40:26 crc kubenswrapper[4980]: I1206 03:40:26.920580 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xhnw8" event={"ID":"3caa51b4-a311-4498-ac9b-10c20989684a","Type":"ContainerDied","Data":"db9e84b373a64698e756936369c3ff5ed646c18669d165181d4f42e254397e0d"} Dec 06 03:40:26 crc kubenswrapper[4980]: I1206 03:40:26.922386 4980 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/56a102c7-f341-4076-89dc-44428e77c164-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 06 03:40:26 crc kubenswrapper[4980]: I1206 03:40:26.922429 4980 scope.go:117] "RemoveContainer" containerID="c9d59ae689acb66f2a450b619034ed36904d9c1cc9adfef732f043b58e582bec" Dec 06 03:40:26 crc kubenswrapper[4980]: I1206 03:40:26.981650 4980 scope.go:117] "RemoveContainer" containerID="d47b4f045fefdb2c7a2a6fe135c197172906e70ebfd6c2ae3c51b3798b9982cc" Dec 06 03:40:27 crc kubenswrapper[4980]: I1206 03:40:27.006149 4980 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-58d94"] Dec 06 03:40:27 crc kubenswrapper[4980]: I1206 03:40:27.013585 4980 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-58d94"] Dec 06 03:40:27 crc kubenswrapper[4980]: I1206 03:40:27.015680 4980 scope.go:117] "RemoveContainer" containerID="db54300afb9839558c5c8d46ba1ba9aa3fac94f6b6e71f3c6511f74fd44c97ad" Dec 06 03:40:27 crc kubenswrapper[4980]: I1206 03:40:27.048505 4980 scope.go:117] "RemoveContainer" containerID="ef2befdf9890e17b7d072571f6b80abbd83962e482172e3dc516ca5c403631c6" Dec 06 03:40:27 crc kubenswrapper[4980]: I1206 03:40:27.056630 4980 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-kp68b"] Dec 06 03:40:27 crc kubenswrapper[4980]: I1206 03:40:27.059087 4980 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-kp68b"] Dec 06 03:40:27 crc kubenswrapper[4980]: I1206 03:40:27.067276 4980 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-m9vl6"] Dec 06 03:40:27 crc kubenswrapper[4980]: I1206 03:40:27.072259 4980 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-m9vl6"] Dec 06 03:40:27 crc kubenswrapper[4980]: I1206 03:40:27.081923 4980 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-2vfzc"] Dec 06 03:40:27 crc kubenswrapper[4980]: I1206 03:40:27.092728 4980 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-2vfzc"] Dec 06 03:40:27 crc kubenswrapper[4980]: I1206 03:40:27.107744 4980 scope.go:117] "RemoveContainer" containerID="739e651d553aa530b056a94ecbc5131776fd3e6197e4a6840089a2c0f03e49b0" Dec 06 03:40:27 crc kubenswrapper[4980]: I1206 03:40:27.129234 4980 scope.go:117] "RemoveContainer" containerID="8aa56292c24f69d83a0939e746f3716f1409df631c637167480a49c129beb2ca" Dec 06 03:40:27 crc kubenswrapper[4980]: I1206 03:40:27.133096 4980 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-gxcqb"] Dec 06 03:40:27 crc kubenswrapper[4980]: I1206 03:40:27.141152 4980 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-gxcqb"] Dec 06 03:40:27 crc kubenswrapper[4980]: I1206 03:40:27.159212 4980 scope.go:117] "RemoveContainer" containerID="1f375c47d1ab2d4e72c1d6e5a9af8ee6f9280fccbd3ce180afdc0ff7b8a980bf" Dec 06 03:40:27 crc kubenswrapper[4980]: I1206 03:40:27.168576 4980 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-xhnw8"] Dec 06 03:40:27 crc kubenswrapper[4980]: I1206 03:40:27.172152 4980 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-xhnw8"] Dec 06 03:40:27 crc kubenswrapper[4980]: I1206 03:40:27.181899 4980 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-spd2v"] Dec 06 03:40:27 crc kubenswrapper[4980]: I1206 03:40:27.200319 4980 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2ce9bc16-3d2a-4feb-82c1-0f42518196cf" path="/var/lib/kubelet/pods/2ce9bc16-3d2a-4feb-82c1-0f42518196cf/volumes" Dec 06 03:40:27 crc kubenswrapper[4980]: I1206 03:40:27.201242 4980 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3caa51b4-a311-4498-ac9b-10c20989684a" path="/var/lib/kubelet/pods/3caa51b4-a311-4498-ac9b-10c20989684a/volumes" Dec 06 03:40:27 crc kubenswrapper[4980]: I1206 03:40:27.210475 4980 scope.go:117] "RemoveContainer" containerID="b6c5027c91c80a97467c1e8c50318fab43b7ec468acda83d6d1f8288d5c58543" Dec 06 03:40:27 crc kubenswrapper[4980]: I1206 03:40:27.211264 4980 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="43112506-9fdd-4b19-a5a5-caf864c09774" path="/var/lib/kubelet/pods/43112506-9fdd-4b19-a5a5-caf864c09774/volumes" Dec 06 03:40:27 crc kubenswrapper[4980]: I1206 03:40:27.212092 4980 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="652a7feb-30e3-4e9f-b1a9-c8ef201c658c" path="/var/lib/kubelet/pods/652a7feb-30e3-4e9f-b1a9-c8ef201c658c/volumes" Dec 06 03:40:27 crc kubenswrapper[4980]: I1206 03:40:27.213818 4980 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="769abfb5-c18a-4cf0-a033-95b3a808867b" path="/var/lib/kubelet/pods/769abfb5-c18a-4cf0-a033-95b3a808867b/volumes" Dec 06 03:40:27 crc kubenswrapper[4980]: I1206 03:40:27.214644 4980 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9c8b3374-77eb-44ad-b270-2263c6c23e28" path="/var/lib/kubelet/pods/9c8b3374-77eb-44ad-b270-2263c6c23e28/volumes" Dec 06 03:40:27 crc kubenswrapper[4980]: I1206 03:40:27.215886 4980 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e965bb7a-1bf0-4574-aae4-2a45c75fb11c" path="/var/lib/kubelet/pods/e965bb7a-1bf0-4574-aae4-2a45c75fb11c/volumes" Dec 06 03:40:27 crc kubenswrapper[4980]: I1206 03:40:27.216555 4980 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-spd2v"] Dec 06 03:40:27 crc kubenswrapper[4980]: I1206 03:40:27.221565 4980 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-bl7hx"] Dec 06 03:40:27 crc kubenswrapper[4980]: I1206 03:40:27.226431 4980 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-bl7hx"] Dec 06 03:40:27 crc kubenswrapper[4980]: I1206 03:40:27.238666 4980 scope.go:117] "RemoveContainer" containerID="fdff05d7789244fc432df351c27d2f5fcb050f2030386cc3a499b53e0c252358" Dec 06 03:40:27 crc kubenswrapper[4980]: I1206 03:40:27.264620 4980 scope.go:117] "RemoveContainer" containerID="6226b5de172f7a25afe31ae6e38b1c241129340606a18fafb76921d2e95f2958" Dec 06 03:40:27 crc kubenswrapper[4980]: I1206 03:40:27.291541 4980 scope.go:117] "RemoveContainer" containerID="10568cc585896cfd066308ee3d8cc363c77bac90cfad8e139177a4794fba172f" Dec 06 03:40:27 crc kubenswrapper[4980]: I1206 03:40:27.309657 4980 scope.go:117] "RemoveContainer" containerID="fdff05d7789244fc432df351c27d2f5fcb050f2030386cc3a499b53e0c252358" Dec 06 03:40:27 crc kubenswrapper[4980]: E1206 03:40:27.310107 4980 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fdff05d7789244fc432df351c27d2f5fcb050f2030386cc3a499b53e0c252358\": container with ID starting with fdff05d7789244fc432df351c27d2f5fcb050f2030386cc3a499b53e0c252358 not found: ID does not exist" containerID="fdff05d7789244fc432df351c27d2f5fcb050f2030386cc3a499b53e0c252358" Dec 06 03:40:27 crc kubenswrapper[4980]: I1206 03:40:27.310275 4980 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fdff05d7789244fc432df351c27d2f5fcb050f2030386cc3a499b53e0c252358"} err="failed to get container status \"fdff05d7789244fc432df351c27d2f5fcb050f2030386cc3a499b53e0c252358\": rpc error: code = NotFound desc = could not find container \"fdff05d7789244fc432df351c27d2f5fcb050f2030386cc3a499b53e0c252358\": container with ID starting with fdff05d7789244fc432df351c27d2f5fcb050f2030386cc3a499b53e0c252358 not found: ID does not exist" Dec 06 03:40:27 crc kubenswrapper[4980]: I1206 03:40:27.310389 4980 scope.go:117] "RemoveContainer" containerID="6226b5de172f7a25afe31ae6e38b1c241129340606a18fafb76921d2e95f2958" Dec 06 03:40:27 crc kubenswrapper[4980]: E1206 03:40:27.310851 4980 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6226b5de172f7a25afe31ae6e38b1c241129340606a18fafb76921d2e95f2958\": container with ID starting with 6226b5de172f7a25afe31ae6e38b1c241129340606a18fafb76921d2e95f2958 not found: ID does not exist" containerID="6226b5de172f7a25afe31ae6e38b1c241129340606a18fafb76921d2e95f2958" Dec 06 03:40:27 crc kubenswrapper[4980]: I1206 03:40:27.310881 4980 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6226b5de172f7a25afe31ae6e38b1c241129340606a18fafb76921d2e95f2958"} err="failed to get container status \"6226b5de172f7a25afe31ae6e38b1c241129340606a18fafb76921d2e95f2958\": rpc error: code = NotFound desc = could not find container \"6226b5de172f7a25afe31ae6e38b1c241129340606a18fafb76921d2e95f2958\": container with ID starting with 6226b5de172f7a25afe31ae6e38b1c241129340606a18fafb76921d2e95f2958 not found: ID does not exist" Dec 06 03:40:27 crc kubenswrapper[4980]: I1206 03:40:27.310901 4980 scope.go:117] "RemoveContainer" containerID="10568cc585896cfd066308ee3d8cc363c77bac90cfad8e139177a4794fba172f" Dec 06 03:40:27 crc kubenswrapper[4980]: E1206 03:40:27.311161 4980 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"10568cc585896cfd066308ee3d8cc363c77bac90cfad8e139177a4794fba172f\": container with ID starting with 10568cc585896cfd066308ee3d8cc363c77bac90cfad8e139177a4794fba172f not found: ID does not exist" containerID="10568cc585896cfd066308ee3d8cc363c77bac90cfad8e139177a4794fba172f" Dec 06 03:40:27 crc kubenswrapper[4980]: I1206 03:40:27.311194 4980 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"10568cc585896cfd066308ee3d8cc363c77bac90cfad8e139177a4794fba172f"} err="failed to get container status \"10568cc585896cfd066308ee3d8cc363c77bac90cfad8e139177a4794fba172f\": rpc error: code = NotFound desc = could not find container \"10568cc585896cfd066308ee3d8cc363c77bac90cfad8e139177a4794fba172f\": container with ID starting with 10568cc585896cfd066308ee3d8cc363c77bac90cfad8e139177a4794fba172f not found: ID does not exist" Dec 06 03:40:27 crc kubenswrapper[4980]: I1206 03:40:27.311212 4980 scope.go:117] "RemoveContainer" containerID="9aa7d598bd03dc4b167375c94a61c6b2efcf6f1430a76effe832f91e012023d1" Dec 06 03:40:27 crc kubenswrapper[4980]: I1206 03:40:27.328009 4980 scope.go:117] "RemoveContainer" containerID="1dacbca162fc3e3abb85ba1b986d4c3a0c7e310dc145b1a6f971a248a74905e8" Dec 06 03:40:27 crc kubenswrapper[4980]: I1206 03:40:27.342559 4980 scope.go:117] "RemoveContainer" containerID="c07bf9bd85f149a01c55d59978da7a3af66f9d797c3c9138c946857d0785f2ac" Dec 06 03:40:27 crc kubenswrapper[4980]: I1206 03:40:27.354854 4980 scope.go:117] "RemoveContainer" containerID="814e5ffce677f097050c8e056ca0ebc60c478a53dbfdeffbc4ad76fabbd8622b" Dec 06 03:40:27 crc kubenswrapper[4980]: I1206 03:40:27.368339 4980 scope.go:117] "RemoveContainer" containerID="beb2e6aabb313ede1d7348363d791410e4e40f483c9f4bfceec700e83315d0b9" Dec 06 03:40:27 crc kubenswrapper[4980]: I1206 03:40:27.383675 4980 scope.go:117] "RemoveContainer" containerID="97406efd5585dbfcd36479b31e92271c541ccac868fb34f708f0309bc89189c4" Dec 06 03:40:27 crc kubenswrapper[4980]: I1206 03:40:27.397770 4980 scope.go:117] "RemoveContainer" containerID="814e5ffce677f097050c8e056ca0ebc60c478a53dbfdeffbc4ad76fabbd8622b" Dec 06 03:40:27 crc kubenswrapper[4980]: E1206 03:40:27.398179 4980 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"814e5ffce677f097050c8e056ca0ebc60c478a53dbfdeffbc4ad76fabbd8622b\": container with ID starting with 814e5ffce677f097050c8e056ca0ebc60c478a53dbfdeffbc4ad76fabbd8622b not found: ID does not exist" containerID="814e5ffce677f097050c8e056ca0ebc60c478a53dbfdeffbc4ad76fabbd8622b" Dec 06 03:40:27 crc kubenswrapper[4980]: I1206 03:40:27.398216 4980 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"814e5ffce677f097050c8e056ca0ebc60c478a53dbfdeffbc4ad76fabbd8622b"} err="failed to get container status \"814e5ffce677f097050c8e056ca0ebc60c478a53dbfdeffbc4ad76fabbd8622b\": rpc error: code = NotFound desc = could not find container \"814e5ffce677f097050c8e056ca0ebc60c478a53dbfdeffbc4ad76fabbd8622b\": container with ID starting with 814e5ffce677f097050c8e056ca0ebc60c478a53dbfdeffbc4ad76fabbd8622b not found: ID does not exist" Dec 06 03:40:27 crc kubenswrapper[4980]: I1206 03:40:27.398245 4980 scope.go:117] "RemoveContainer" containerID="beb2e6aabb313ede1d7348363d791410e4e40f483c9f4bfceec700e83315d0b9" Dec 06 03:40:27 crc kubenswrapper[4980]: E1206 03:40:27.398499 4980 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"beb2e6aabb313ede1d7348363d791410e4e40f483c9f4bfceec700e83315d0b9\": container with ID starting with beb2e6aabb313ede1d7348363d791410e4e40f483c9f4bfceec700e83315d0b9 not found: ID does not exist" containerID="beb2e6aabb313ede1d7348363d791410e4e40f483c9f4bfceec700e83315d0b9" Dec 06 03:40:27 crc kubenswrapper[4980]: I1206 03:40:27.398540 4980 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"beb2e6aabb313ede1d7348363d791410e4e40f483c9f4bfceec700e83315d0b9"} err="failed to get container status \"beb2e6aabb313ede1d7348363d791410e4e40f483c9f4bfceec700e83315d0b9\": rpc error: code = NotFound desc = could not find container \"beb2e6aabb313ede1d7348363d791410e4e40f483c9f4bfceec700e83315d0b9\": container with ID starting with beb2e6aabb313ede1d7348363d791410e4e40f483c9f4bfceec700e83315d0b9 not found: ID does not exist" Dec 06 03:40:27 crc kubenswrapper[4980]: I1206 03:40:27.398557 4980 scope.go:117] "RemoveContainer" containerID="97406efd5585dbfcd36479b31e92271c541ccac868fb34f708f0309bc89189c4" Dec 06 03:40:27 crc kubenswrapper[4980]: E1206 03:40:27.399002 4980 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"97406efd5585dbfcd36479b31e92271c541ccac868fb34f708f0309bc89189c4\": container with ID starting with 97406efd5585dbfcd36479b31e92271c541ccac868fb34f708f0309bc89189c4 not found: ID does not exist" containerID="97406efd5585dbfcd36479b31e92271c541ccac868fb34f708f0309bc89189c4" Dec 06 03:40:27 crc kubenswrapper[4980]: I1206 03:40:27.399025 4980 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"97406efd5585dbfcd36479b31e92271c541ccac868fb34f708f0309bc89189c4"} err="failed to get container status \"97406efd5585dbfcd36479b31e92271c541ccac868fb34f708f0309bc89189c4\": rpc error: code = NotFound desc = could not find container \"97406efd5585dbfcd36479b31e92271c541ccac868fb34f708f0309bc89189c4\": container with ID starting with 97406efd5585dbfcd36479b31e92271c541ccac868fb34f708f0309bc89189c4 not found: ID does not exist" Dec 06 03:40:27 crc kubenswrapper[4980]: I1206 03:40:27.399039 4980 scope.go:117] "RemoveContainer" containerID="f0efc2ce0c1863787e77589a8a0afc2f9e4fdb3ce974b5bcac67705360f28d24" Dec 06 03:40:27 crc kubenswrapper[4980]: I1206 03:40:27.410610 4980 scope.go:117] "RemoveContainer" containerID="ffbb4c8c6980bde657d7596d37a800b08dd5ac764045bebabcafe2c50bd30ac6" Dec 06 03:40:27 crc kubenswrapper[4980]: I1206 03:40:27.423683 4980 scope.go:117] "RemoveContainer" containerID="a3b66bc131bf66af908ba3cf9ed3c9d469b481a1db21f645f7785f7051d84f9f" Dec 06 03:40:27 crc kubenswrapper[4980]: I1206 03:40:27.443408 4980 scope.go:117] "RemoveContainer" containerID="e5e0adc7836a520047580410eb93beb81ac6a2dad1bc364b0c5bcf599f13cf86" Dec 06 03:40:27 crc kubenswrapper[4980]: I1206 03:40:27.458567 4980 scope.go:117] "RemoveContainer" containerID="3fd75664d84b78f5765b694a48cb11d62122b84f14e3906a4eafb96f5bd2011c" Dec 06 03:40:27 crc kubenswrapper[4980]: I1206 03:40:27.470262 4980 scope.go:117] "RemoveContainer" containerID="863754232f8ecce26db582347b737e0fae6c66c705dcb1ed803f6da176d756e1" Dec 06 03:40:27 crc kubenswrapper[4980]: I1206 03:40:27.482996 4980 scope.go:117] "RemoveContainer" containerID="e5e0adc7836a520047580410eb93beb81ac6a2dad1bc364b0c5bcf599f13cf86" Dec 06 03:40:27 crc kubenswrapper[4980]: E1206 03:40:27.483756 4980 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e5e0adc7836a520047580410eb93beb81ac6a2dad1bc364b0c5bcf599f13cf86\": container with ID starting with e5e0adc7836a520047580410eb93beb81ac6a2dad1bc364b0c5bcf599f13cf86 not found: ID does not exist" containerID="e5e0adc7836a520047580410eb93beb81ac6a2dad1bc364b0c5bcf599f13cf86" Dec 06 03:40:27 crc kubenswrapper[4980]: I1206 03:40:27.483799 4980 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e5e0adc7836a520047580410eb93beb81ac6a2dad1bc364b0c5bcf599f13cf86"} err="failed to get container status \"e5e0adc7836a520047580410eb93beb81ac6a2dad1bc364b0c5bcf599f13cf86\": rpc error: code = NotFound desc = could not find container \"e5e0adc7836a520047580410eb93beb81ac6a2dad1bc364b0c5bcf599f13cf86\": container with ID starting with e5e0adc7836a520047580410eb93beb81ac6a2dad1bc364b0c5bcf599f13cf86 not found: ID does not exist" Dec 06 03:40:27 crc kubenswrapper[4980]: I1206 03:40:27.483831 4980 scope.go:117] "RemoveContainer" containerID="3fd75664d84b78f5765b694a48cb11d62122b84f14e3906a4eafb96f5bd2011c" Dec 06 03:40:27 crc kubenswrapper[4980]: E1206 03:40:27.484231 4980 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3fd75664d84b78f5765b694a48cb11d62122b84f14e3906a4eafb96f5bd2011c\": container with ID starting with 3fd75664d84b78f5765b694a48cb11d62122b84f14e3906a4eafb96f5bd2011c not found: ID does not exist" containerID="3fd75664d84b78f5765b694a48cb11d62122b84f14e3906a4eafb96f5bd2011c" Dec 06 03:40:27 crc kubenswrapper[4980]: I1206 03:40:27.484338 4980 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3fd75664d84b78f5765b694a48cb11d62122b84f14e3906a4eafb96f5bd2011c"} err="failed to get container status \"3fd75664d84b78f5765b694a48cb11d62122b84f14e3906a4eafb96f5bd2011c\": rpc error: code = NotFound desc = could not find container \"3fd75664d84b78f5765b694a48cb11d62122b84f14e3906a4eafb96f5bd2011c\": container with ID starting with 3fd75664d84b78f5765b694a48cb11d62122b84f14e3906a4eafb96f5bd2011c not found: ID does not exist" Dec 06 03:40:27 crc kubenswrapper[4980]: I1206 03:40:27.484420 4980 scope.go:117] "RemoveContainer" containerID="863754232f8ecce26db582347b737e0fae6c66c705dcb1ed803f6da176d756e1" Dec 06 03:40:27 crc kubenswrapper[4980]: E1206 03:40:27.484739 4980 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"863754232f8ecce26db582347b737e0fae6c66c705dcb1ed803f6da176d756e1\": container with ID starting with 863754232f8ecce26db582347b737e0fae6c66c705dcb1ed803f6da176d756e1 not found: ID does not exist" containerID="863754232f8ecce26db582347b737e0fae6c66c705dcb1ed803f6da176d756e1" Dec 06 03:40:27 crc kubenswrapper[4980]: I1206 03:40:27.484768 4980 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"863754232f8ecce26db582347b737e0fae6c66c705dcb1ed803f6da176d756e1"} err="failed to get container status \"863754232f8ecce26db582347b737e0fae6c66c705dcb1ed803f6da176d756e1\": rpc error: code = NotFound desc = could not find container \"863754232f8ecce26db582347b737e0fae6c66c705dcb1ed803f6da176d756e1\": container with ID starting with 863754232f8ecce26db582347b737e0fae6c66c705dcb1ed803f6da176d756e1 not found: ID does not exist" Dec 06 03:40:27 crc kubenswrapper[4980]: I1206 03:40:27.939808 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-rhhmp" event={"ID":"34c9fd26-4164-4267-8dce-7f487139e741","Type":"ContainerStarted","Data":"4a81b88f9eb14d54f6f9d3f9f668da4d43d1155662f9e6ead2bccf42b502492f"} Dec 06 03:40:27 crc kubenswrapper[4980]: I1206 03:40:27.941567 4980 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-rhhmp" Dec 06 03:40:27 crc kubenswrapper[4980]: I1206 03:40:27.941669 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-rhhmp" event={"ID":"34c9fd26-4164-4267-8dce-7f487139e741","Type":"ContainerStarted","Data":"0760dd57d1a64eeb554bda3abcf1ce38a2a1a80403e88f06addf66b2884eb7bd"} Dec 06 03:40:27 crc kubenswrapper[4980]: I1206 03:40:27.944044 4980 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-rhhmp" Dec 06 03:40:27 crc kubenswrapper[4980]: I1206 03:40:27.984862 4980 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-rhhmp" podStartSLOduration=2.98483205 podStartE2EDuration="2.98483205s" podCreationTimestamp="2025-12-06 03:40:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 03:40:27.963923424 +0000 UTC m=+387.208674735" watchObservedRunningTime="2025-12-06 03:40:27.98483205 +0000 UTC m=+387.229583351" Dec 06 03:40:29 crc kubenswrapper[4980]: I1206 03:40:29.191312 4980 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0a45bc54-be9c-4c4f-9cae-dc369d25be33" path="/var/lib/kubelet/pods/0a45bc54-be9c-4c4f-9cae-dc369d25be33/volumes" Dec 06 03:40:29 crc kubenswrapper[4980]: I1206 03:40:29.192054 4980 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="56a102c7-f341-4076-89dc-44428e77c164" path="/var/lib/kubelet/pods/56a102c7-f341-4076-89dc-44428e77c164/volumes" Dec 06 03:40:37 crc kubenswrapper[4980]: I1206 03:40:37.014184 4980 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-x8x8b"] Dec 06 03:40:37 crc kubenswrapper[4980]: E1206 03:40:37.015163 4980 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="56a102c7-f341-4076-89dc-44428e77c164" containerName="registry-server" Dec 06 03:40:37 crc kubenswrapper[4980]: I1206 03:40:37.015195 4980 state_mem.go:107] "Deleted CPUSet assignment" podUID="56a102c7-f341-4076-89dc-44428e77c164" containerName="registry-server" Dec 06 03:40:37 crc kubenswrapper[4980]: E1206 03:40:37.015218 4980 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e965bb7a-1bf0-4574-aae4-2a45c75fb11c" containerName="registry-server" Dec 06 03:40:37 crc kubenswrapper[4980]: I1206 03:40:37.015229 4980 state_mem.go:107] "Deleted CPUSet assignment" podUID="e965bb7a-1bf0-4574-aae4-2a45c75fb11c" containerName="registry-server" Dec 06 03:40:37 crc kubenswrapper[4980]: E1206 03:40:37.015243 4980 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="652a7feb-30e3-4e9f-b1a9-c8ef201c658c" containerName="extract-content" Dec 06 03:40:37 crc kubenswrapper[4980]: I1206 03:40:37.015253 4980 state_mem.go:107] "Deleted CPUSet assignment" podUID="652a7feb-30e3-4e9f-b1a9-c8ef201c658c" containerName="extract-content" Dec 06 03:40:37 crc kubenswrapper[4980]: E1206 03:40:37.015270 4980 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="56a102c7-f341-4076-89dc-44428e77c164" containerName="extract-content" Dec 06 03:40:37 crc kubenswrapper[4980]: I1206 03:40:37.015279 4980 state_mem.go:107] "Deleted CPUSet assignment" podUID="56a102c7-f341-4076-89dc-44428e77c164" containerName="extract-content" Dec 06 03:40:37 crc kubenswrapper[4980]: E1206 03:40:37.015295 4980 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3caa51b4-a311-4498-ac9b-10c20989684a" containerName="extract-utilities" Dec 06 03:40:37 crc kubenswrapper[4980]: I1206 03:40:37.015304 4980 state_mem.go:107] "Deleted CPUSet assignment" podUID="3caa51b4-a311-4498-ac9b-10c20989684a" containerName="extract-utilities" Dec 06 03:40:37 crc kubenswrapper[4980]: E1206 03:40:37.015312 4980 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9c8b3374-77eb-44ad-b270-2263c6c23e28" containerName="registry-server" Dec 06 03:40:37 crc kubenswrapper[4980]: I1206 03:40:37.015320 4980 state_mem.go:107] "Deleted CPUSet assignment" podUID="9c8b3374-77eb-44ad-b270-2263c6c23e28" containerName="registry-server" Dec 06 03:40:37 crc kubenswrapper[4980]: E1206 03:40:37.015331 4980 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="769abfb5-c18a-4cf0-a033-95b3a808867b" containerName="extract-utilities" Dec 06 03:40:37 crc kubenswrapper[4980]: I1206 03:40:37.015338 4980 state_mem.go:107] "Deleted CPUSet assignment" podUID="769abfb5-c18a-4cf0-a033-95b3a808867b" containerName="extract-utilities" Dec 06 03:40:37 crc kubenswrapper[4980]: E1206 03:40:37.015352 4980 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0a45bc54-be9c-4c4f-9cae-dc369d25be33" containerName="marketplace-operator" Dec 06 03:40:37 crc kubenswrapper[4980]: I1206 03:40:37.015359 4980 state_mem.go:107] "Deleted CPUSet assignment" podUID="0a45bc54-be9c-4c4f-9cae-dc369d25be33" containerName="marketplace-operator" Dec 06 03:40:37 crc kubenswrapper[4980]: E1206 03:40:37.015367 4980 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2ce9bc16-3d2a-4feb-82c1-0f42518196cf" containerName="extract-content" Dec 06 03:40:37 crc kubenswrapper[4980]: I1206 03:40:37.015375 4980 state_mem.go:107] "Deleted CPUSet assignment" podUID="2ce9bc16-3d2a-4feb-82c1-0f42518196cf" containerName="extract-content" Dec 06 03:40:37 crc kubenswrapper[4980]: E1206 03:40:37.015388 4980 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3caa51b4-a311-4498-ac9b-10c20989684a" containerName="extract-content" Dec 06 03:40:37 crc kubenswrapper[4980]: I1206 03:40:37.015395 4980 state_mem.go:107] "Deleted CPUSet assignment" podUID="3caa51b4-a311-4498-ac9b-10c20989684a" containerName="extract-content" Dec 06 03:40:37 crc kubenswrapper[4980]: E1206 03:40:37.015407 4980 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e965bb7a-1bf0-4574-aae4-2a45c75fb11c" containerName="extract-content" Dec 06 03:40:37 crc kubenswrapper[4980]: I1206 03:40:37.015416 4980 state_mem.go:107] "Deleted CPUSet assignment" podUID="e965bb7a-1bf0-4574-aae4-2a45c75fb11c" containerName="extract-content" Dec 06 03:40:37 crc kubenswrapper[4980]: E1206 03:40:37.015433 4980 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2ce9bc16-3d2a-4feb-82c1-0f42518196cf" containerName="extract-utilities" Dec 06 03:40:37 crc kubenswrapper[4980]: I1206 03:40:37.015445 4980 state_mem.go:107] "Deleted CPUSet assignment" podUID="2ce9bc16-3d2a-4feb-82c1-0f42518196cf" containerName="extract-utilities" Dec 06 03:40:37 crc kubenswrapper[4980]: E1206 03:40:37.015458 4980 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="43112506-9fdd-4b19-a5a5-caf864c09774" containerName="extract-content" Dec 06 03:40:37 crc kubenswrapper[4980]: I1206 03:40:37.015468 4980 state_mem.go:107] "Deleted CPUSet assignment" podUID="43112506-9fdd-4b19-a5a5-caf864c09774" containerName="extract-content" Dec 06 03:40:37 crc kubenswrapper[4980]: E1206 03:40:37.015477 4980 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="43112506-9fdd-4b19-a5a5-caf864c09774" containerName="extract-utilities" Dec 06 03:40:37 crc kubenswrapper[4980]: I1206 03:40:37.015485 4980 state_mem.go:107] "Deleted CPUSet assignment" podUID="43112506-9fdd-4b19-a5a5-caf864c09774" containerName="extract-utilities" Dec 06 03:40:37 crc kubenswrapper[4980]: E1206 03:40:37.015497 4980 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="43112506-9fdd-4b19-a5a5-caf864c09774" containerName="registry-server" Dec 06 03:40:37 crc kubenswrapper[4980]: I1206 03:40:37.015505 4980 state_mem.go:107] "Deleted CPUSet assignment" podUID="43112506-9fdd-4b19-a5a5-caf864c09774" containerName="registry-server" Dec 06 03:40:37 crc kubenswrapper[4980]: E1206 03:40:37.015538 4980 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e965bb7a-1bf0-4574-aae4-2a45c75fb11c" containerName="extract-utilities" Dec 06 03:40:37 crc kubenswrapper[4980]: I1206 03:40:37.015548 4980 state_mem.go:107] "Deleted CPUSet assignment" podUID="e965bb7a-1bf0-4574-aae4-2a45c75fb11c" containerName="extract-utilities" Dec 06 03:40:37 crc kubenswrapper[4980]: E1206 03:40:37.015558 4980 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="652a7feb-30e3-4e9f-b1a9-c8ef201c658c" containerName="registry-server" Dec 06 03:40:37 crc kubenswrapper[4980]: I1206 03:40:37.015567 4980 state_mem.go:107] "Deleted CPUSet assignment" podUID="652a7feb-30e3-4e9f-b1a9-c8ef201c658c" containerName="registry-server" Dec 06 03:40:37 crc kubenswrapper[4980]: E1206 03:40:37.015578 4980 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="769abfb5-c18a-4cf0-a033-95b3a808867b" containerName="extract-content" Dec 06 03:40:37 crc kubenswrapper[4980]: I1206 03:40:37.015586 4980 state_mem.go:107] "Deleted CPUSet assignment" podUID="769abfb5-c18a-4cf0-a033-95b3a808867b" containerName="extract-content" Dec 06 03:40:37 crc kubenswrapper[4980]: E1206 03:40:37.015596 4980 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9c8b3374-77eb-44ad-b270-2263c6c23e28" containerName="extract-utilities" Dec 06 03:40:37 crc kubenswrapper[4980]: I1206 03:40:37.015603 4980 state_mem.go:107] "Deleted CPUSet assignment" podUID="9c8b3374-77eb-44ad-b270-2263c6c23e28" containerName="extract-utilities" Dec 06 03:40:37 crc kubenswrapper[4980]: E1206 03:40:37.015613 4980 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="652a7feb-30e3-4e9f-b1a9-c8ef201c658c" containerName="extract-utilities" Dec 06 03:40:37 crc kubenswrapper[4980]: I1206 03:40:37.015622 4980 state_mem.go:107] "Deleted CPUSet assignment" podUID="652a7feb-30e3-4e9f-b1a9-c8ef201c658c" containerName="extract-utilities" Dec 06 03:40:37 crc kubenswrapper[4980]: E1206 03:40:37.015633 4980 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3caa51b4-a311-4498-ac9b-10c20989684a" containerName="registry-server" Dec 06 03:40:37 crc kubenswrapper[4980]: I1206 03:40:37.015642 4980 state_mem.go:107] "Deleted CPUSet assignment" podUID="3caa51b4-a311-4498-ac9b-10c20989684a" containerName="registry-server" Dec 06 03:40:37 crc kubenswrapper[4980]: E1206 03:40:37.015653 4980 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="769abfb5-c18a-4cf0-a033-95b3a808867b" containerName="registry-server" Dec 06 03:40:37 crc kubenswrapper[4980]: I1206 03:40:37.015662 4980 state_mem.go:107] "Deleted CPUSet assignment" podUID="769abfb5-c18a-4cf0-a033-95b3a808867b" containerName="registry-server" Dec 06 03:40:37 crc kubenswrapper[4980]: E1206 03:40:37.015677 4980 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9c8b3374-77eb-44ad-b270-2263c6c23e28" containerName="extract-content" Dec 06 03:40:37 crc kubenswrapper[4980]: I1206 03:40:37.015686 4980 state_mem.go:107] "Deleted CPUSet assignment" podUID="9c8b3374-77eb-44ad-b270-2263c6c23e28" containerName="extract-content" Dec 06 03:40:37 crc kubenswrapper[4980]: E1206 03:40:37.015700 4980 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="56a102c7-f341-4076-89dc-44428e77c164" containerName="extract-utilities" Dec 06 03:40:37 crc kubenswrapper[4980]: I1206 03:40:37.015710 4980 state_mem.go:107] "Deleted CPUSet assignment" podUID="56a102c7-f341-4076-89dc-44428e77c164" containerName="extract-utilities" Dec 06 03:40:37 crc kubenswrapper[4980]: E1206 03:40:37.015721 4980 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2ce9bc16-3d2a-4feb-82c1-0f42518196cf" containerName="registry-server" Dec 06 03:40:37 crc kubenswrapper[4980]: I1206 03:40:37.015730 4980 state_mem.go:107] "Deleted CPUSet assignment" podUID="2ce9bc16-3d2a-4feb-82c1-0f42518196cf" containerName="registry-server" Dec 06 03:40:37 crc kubenswrapper[4980]: I1206 03:40:37.015847 4980 memory_manager.go:354] "RemoveStaleState removing state" podUID="43112506-9fdd-4b19-a5a5-caf864c09774" containerName="registry-server" Dec 06 03:40:37 crc kubenswrapper[4980]: I1206 03:40:37.015860 4980 memory_manager.go:354] "RemoveStaleState removing state" podUID="9c8b3374-77eb-44ad-b270-2263c6c23e28" containerName="registry-server" Dec 06 03:40:37 crc kubenswrapper[4980]: I1206 03:40:37.015872 4980 memory_manager.go:354] "RemoveStaleState removing state" podUID="3caa51b4-a311-4498-ac9b-10c20989684a" containerName="registry-server" Dec 06 03:40:37 crc kubenswrapper[4980]: I1206 03:40:37.015883 4980 memory_manager.go:354] "RemoveStaleState removing state" podUID="2ce9bc16-3d2a-4feb-82c1-0f42518196cf" containerName="registry-server" Dec 06 03:40:37 crc kubenswrapper[4980]: I1206 03:40:37.015893 4980 memory_manager.go:354] "RemoveStaleState removing state" podUID="769abfb5-c18a-4cf0-a033-95b3a808867b" containerName="registry-server" Dec 06 03:40:37 crc kubenswrapper[4980]: I1206 03:40:37.015902 4980 memory_manager.go:354] "RemoveStaleState removing state" podUID="652a7feb-30e3-4e9f-b1a9-c8ef201c658c" containerName="registry-server" Dec 06 03:40:37 crc kubenswrapper[4980]: I1206 03:40:37.015911 4980 memory_manager.go:354] "RemoveStaleState removing state" podUID="0a45bc54-be9c-4c4f-9cae-dc369d25be33" containerName="marketplace-operator" Dec 06 03:40:37 crc kubenswrapper[4980]: I1206 03:40:37.015921 4980 memory_manager.go:354] "RemoveStaleState removing state" podUID="56a102c7-f341-4076-89dc-44428e77c164" containerName="registry-server" Dec 06 03:40:37 crc kubenswrapper[4980]: I1206 03:40:37.015934 4980 memory_manager.go:354] "RemoveStaleState removing state" podUID="e965bb7a-1bf0-4574-aae4-2a45c75fb11c" containerName="registry-server" Dec 06 03:40:37 crc kubenswrapper[4980]: I1206 03:40:37.017020 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-x8x8b" Dec 06 03:40:37 crc kubenswrapper[4980]: I1206 03:40:37.019119 4980 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Dec 06 03:40:37 crc kubenswrapper[4980]: I1206 03:40:37.025328 4980 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-x8x8b"] Dec 06 03:40:37 crc kubenswrapper[4980]: I1206 03:40:37.083752 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b6205a6a-fc2b-4a1e-aefd-b86867f3478d-catalog-content\") pod \"community-operators-x8x8b\" (UID: \"b6205a6a-fc2b-4a1e-aefd-b86867f3478d\") " pod="openshift-marketplace/community-operators-x8x8b" Dec 06 03:40:37 crc kubenswrapper[4980]: I1206 03:40:37.083806 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b6205a6a-fc2b-4a1e-aefd-b86867f3478d-utilities\") pod \"community-operators-x8x8b\" (UID: \"b6205a6a-fc2b-4a1e-aefd-b86867f3478d\") " pod="openshift-marketplace/community-operators-x8x8b" Dec 06 03:40:37 crc kubenswrapper[4980]: I1206 03:40:37.083865 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bdtmr\" (UniqueName: \"kubernetes.io/projected/b6205a6a-fc2b-4a1e-aefd-b86867f3478d-kube-api-access-bdtmr\") pod \"community-operators-x8x8b\" (UID: \"b6205a6a-fc2b-4a1e-aefd-b86867f3478d\") " pod="openshift-marketplace/community-operators-x8x8b" Dec 06 03:40:37 crc kubenswrapper[4980]: I1206 03:40:37.185218 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bdtmr\" (UniqueName: \"kubernetes.io/projected/b6205a6a-fc2b-4a1e-aefd-b86867f3478d-kube-api-access-bdtmr\") pod \"community-operators-x8x8b\" (UID: \"b6205a6a-fc2b-4a1e-aefd-b86867f3478d\") " pod="openshift-marketplace/community-operators-x8x8b" Dec 06 03:40:37 crc kubenswrapper[4980]: I1206 03:40:37.185304 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b6205a6a-fc2b-4a1e-aefd-b86867f3478d-catalog-content\") pod \"community-operators-x8x8b\" (UID: \"b6205a6a-fc2b-4a1e-aefd-b86867f3478d\") " pod="openshift-marketplace/community-operators-x8x8b" Dec 06 03:40:37 crc kubenswrapper[4980]: I1206 03:40:37.185340 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b6205a6a-fc2b-4a1e-aefd-b86867f3478d-utilities\") pod \"community-operators-x8x8b\" (UID: \"b6205a6a-fc2b-4a1e-aefd-b86867f3478d\") " pod="openshift-marketplace/community-operators-x8x8b" Dec 06 03:40:37 crc kubenswrapper[4980]: I1206 03:40:37.185987 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b6205a6a-fc2b-4a1e-aefd-b86867f3478d-catalog-content\") pod \"community-operators-x8x8b\" (UID: \"b6205a6a-fc2b-4a1e-aefd-b86867f3478d\") " pod="openshift-marketplace/community-operators-x8x8b" Dec 06 03:40:37 crc kubenswrapper[4980]: I1206 03:40:37.185984 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b6205a6a-fc2b-4a1e-aefd-b86867f3478d-utilities\") pod \"community-operators-x8x8b\" (UID: \"b6205a6a-fc2b-4a1e-aefd-b86867f3478d\") " pod="openshift-marketplace/community-operators-x8x8b" Dec 06 03:40:37 crc kubenswrapper[4980]: I1206 03:40:37.206209 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bdtmr\" (UniqueName: \"kubernetes.io/projected/b6205a6a-fc2b-4a1e-aefd-b86867f3478d-kube-api-access-bdtmr\") pod \"community-operators-x8x8b\" (UID: \"b6205a6a-fc2b-4a1e-aefd-b86867f3478d\") " pod="openshift-marketplace/community-operators-x8x8b" Dec 06 03:40:37 crc kubenswrapper[4980]: I1206 03:40:37.231012 4980 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-d2hwb"] Dec 06 03:40:37 crc kubenswrapper[4980]: I1206 03:40:37.233097 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-d2hwb" Dec 06 03:40:37 crc kubenswrapper[4980]: I1206 03:40:37.237409 4980 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Dec 06 03:40:37 crc kubenswrapper[4980]: I1206 03:40:37.242353 4980 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-d2hwb"] Dec 06 03:40:37 crc kubenswrapper[4980]: I1206 03:40:37.286411 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6b108c98-9b6c-48f5-a6a9-2db36a4e5283-catalog-content\") pod \"certified-operators-d2hwb\" (UID: \"6b108c98-9b6c-48f5-a6a9-2db36a4e5283\") " pod="openshift-marketplace/certified-operators-d2hwb" Dec 06 03:40:37 crc kubenswrapper[4980]: I1206 03:40:37.286794 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6b108c98-9b6c-48f5-a6a9-2db36a4e5283-utilities\") pod \"certified-operators-d2hwb\" (UID: \"6b108c98-9b6c-48f5-a6a9-2db36a4e5283\") " pod="openshift-marketplace/certified-operators-d2hwb" Dec 06 03:40:37 crc kubenswrapper[4980]: I1206 03:40:37.287181 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zq8h7\" (UniqueName: \"kubernetes.io/projected/6b108c98-9b6c-48f5-a6a9-2db36a4e5283-kube-api-access-zq8h7\") pod \"certified-operators-d2hwb\" (UID: \"6b108c98-9b6c-48f5-a6a9-2db36a4e5283\") " pod="openshift-marketplace/certified-operators-d2hwb" Dec 06 03:40:37 crc kubenswrapper[4980]: I1206 03:40:37.337323 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-x8x8b" Dec 06 03:40:37 crc kubenswrapper[4980]: I1206 03:40:37.388176 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zq8h7\" (UniqueName: \"kubernetes.io/projected/6b108c98-9b6c-48f5-a6a9-2db36a4e5283-kube-api-access-zq8h7\") pod \"certified-operators-d2hwb\" (UID: \"6b108c98-9b6c-48f5-a6a9-2db36a4e5283\") " pod="openshift-marketplace/certified-operators-d2hwb" Dec 06 03:40:37 crc kubenswrapper[4980]: I1206 03:40:37.388261 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6b108c98-9b6c-48f5-a6a9-2db36a4e5283-catalog-content\") pod \"certified-operators-d2hwb\" (UID: \"6b108c98-9b6c-48f5-a6a9-2db36a4e5283\") " pod="openshift-marketplace/certified-operators-d2hwb" Dec 06 03:40:37 crc kubenswrapper[4980]: I1206 03:40:37.388300 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6b108c98-9b6c-48f5-a6a9-2db36a4e5283-utilities\") pod \"certified-operators-d2hwb\" (UID: \"6b108c98-9b6c-48f5-a6a9-2db36a4e5283\") " pod="openshift-marketplace/certified-operators-d2hwb" Dec 06 03:40:37 crc kubenswrapper[4980]: I1206 03:40:37.389237 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6b108c98-9b6c-48f5-a6a9-2db36a4e5283-utilities\") pod \"certified-operators-d2hwb\" (UID: \"6b108c98-9b6c-48f5-a6a9-2db36a4e5283\") " pod="openshift-marketplace/certified-operators-d2hwb" Dec 06 03:40:37 crc kubenswrapper[4980]: I1206 03:40:37.391368 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6b108c98-9b6c-48f5-a6a9-2db36a4e5283-catalog-content\") pod \"certified-operators-d2hwb\" (UID: \"6b108c98-9b6c-48f5-a6a9-2db36a4e5283\") " pod="openshift-marketplace/certified-operators-d2hwb" Dec 06 03:40:37 crc kubenswrapper[4980]: I1206 03:40:37.408590 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zq8h7\" (UniqueName: \"kubernetes.io/projected/6b108c98-9b6c-48f5-a6a9-2db36a4e5283-kube-api-access-zq8h7\") pod \"certified-operators-d2hwb\" (UID: \"6b108c98-9b6c-48f5-a6a9-2db36a4e5283\") " pod="openshift-marketplace/certified-operators-d2hwb" Dec 06 03:40:37 crc kubenswrapper[4980]: I1206 03:40:37.567588 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-d2hwb" Dec 06 03:40:37 crc kubenswrapper[4980]: I1206 03:40:37.752488 4980 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-x8x8b"] Dec 06 03:40:37 crc kubenswrapper[4980]: W1206 03:40:37.757689 4980 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb6205a6a_fc2b_4a1e_aefd_b86867f3478d.slice/crio-c2d65fe36224c54c8771a3d38ab1d3e119ddb9622e06029c1d6cb08349dce617 WatchSource:0}: Error finding container c2d65fe36224c54c8771a3d38ab1d3e119ddb9622e06029c1d6cb08349dce617: Status 404 returned error can't find the container with id c2d65fe36224c54c8771a3d38ab1d3e119ddb9622e06029c1d6cb08349dce617 Dec 06 03:40:37 crc kubenswrapper[4980]: I1206 03:40:37.979955 4980 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-d2hwb"] Dec 06 03:40:38 crc kubenswrapper[4980]: I1206 03:40:38.011434 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-d2hwb" event={"ID":"6b108c98-9b6c-48f5-a6a9-2db36a4e5283","Type":"ContainerStarted","Data":"6cfd6774fa2a4ea80ddc7a2502e15ac3f0d0170a24e19f4d331d8b6573dcc4b5"} Dec 06 03:40:38 crc kubenswrapper[4980]: I1206 03:40:38.013231 4980 generic.go:334] "Generic (PLEG): container finished" podID="b6205a6a-fc2b-4a1e-aefd-b86867f3478d" containerID="8e92f4a4a6470e7e8f9ac8bf0ed7cf141ae4419d9083bcf525f9a81a01582ca2" exitCode=0 Dec 06 03:40:38 crc kubenswrapper[4980]: I1206 03:40:38.013274 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-x8x8b" event={"ID":"b6205a6a-fc2b-4a1e-aefd-b86867f3478d","Type":"ContainerDied","Data":"8e92f4a4a6470e7e8f9ac8bf0ed7cf141ae4419d9083bcf525f9a81a01582ca2"} Dec 06 03:40:38 crc kubenswrapper[4980]: I1206 03:40:38.013296 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-x8x8b" event={"ID":"b6205a6a-fc2b-4a1e-aefd-b86867f3478d","Type":"ContainerStarted","Data":"c2d65fe36224c54c8771a3d38ab1d3e119ddb9622e06029c1d6cb08349dce617"} Dec 06 03:40:39 crc kubenswrapper[4980]: I1206 03:40:39.019275 4980 generic.go:334] "Generic (PLEG): container finished" podID="6b108c98-9b6c-48f5-a6a9-2db36a4e5283" containerID="31057024bf7578e59c873141eb7404ce2cd00d32e678445af1fb5c1731ce41b4" exitCode=0 Dec 06 03:40:39 crc kubenswrapper[4980]: I1206 03:40:39.019324 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-d2hwb" event={"ID":"6b108c98-9b6c-48f5-a6a9-2db36a4e5283","Type":"ContainerDied","Data":"31057024bf7578e59c873141eb7404ce2cd00d32e678445af1fb5c1731ce41b4"} Dec 06 03:40:39 crc kubenswrapper[4980]: I1206 03:40:39.408639 4980 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-tgld8"] Dec 06 03:40:39 crc kubenswrapper[4980]: I1206 03:40:39.410265 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-tgld8" Dec 06 03:40:39 crc kubenswrapper[4980]: I1206 03:40:39.413834 4980 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Dec 06 03:40:39 crc kubenswrapper[4980]: I1206 03:40:39.436123 4980 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-tgld8"] Dec 06 03:40:39 crc kubenswrapper[4980]: I1206 03:40:39.537924 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2392c238-d47c-472f-b8f3-cf6ad2720454-catalog-content\") pod \"redhat-marketplace-tgld8\" (UID: \"2392c238-d47c-472f-b8f3-cf6ad2720454\") " pod="openshift-marketplace/redhat-marketplace-tgld8" Dec 06 03:40:39 crc kubenswrapper[4980]: I1206 03:40:39.537978 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bcmq5\" (UniqueName: \"kubernetes.io/projected/2392c238-d47c-472f-b8f3-cf6ad2720454-kube-api-access-bcmq5\") pod \"redhat-marketplace-tgld8\" (UID: \"2392c238-d47c-472f-b8f3-cf6ad2720454\") " pod="openshift-marketplace/redhat-marketplace-tgld8" Dec 06 03:40:39 crc kubenswrapper[4980]: I1206 03:40:39.538020 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2392c238-d47c-472f-b8f3-cf6ad2720454-utilities\") pod \"redhat-marketplace-tgld8\" (UID: \"2392c238-d47c-472f-b8f3-cf6ad2720454\") " pod="openshift-marketplace/redhat-marketplace-tgld8" Dec 06 03:40:39 crc kubenswrapper[4980]: I1206 03:40:39.604064 4980 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-kfc6d"] Dec 06 03:40:39 crc kubenswrapper[4980]: I1206 03:40:39.605030 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-kfc6d" Dec 06 03:40:39 crc kubenswrapper[4980]: I1206 03:40:39.607071 4980 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Dec 06 03:40:39 crc kubenswrapper[4980]: I1206 03:40:39.616105 4980 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-kfc6d"] Dec 06 03:40:39 crc kubenswrapper[4980]: I1206 03:40:39.640036 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/aca9683f-a0d4-46f8-aacb-f2e45896458c-catalog-content\") pod \"redhat-operators-kfc6d\" (UID: \"aca9683f-a0d4-46f8-aacb-f2e45896458c\") " pod="openshift-marketplace/redhat-operators-kfc6d" Dec 06 03:40:39 crc kubenswrapper[4980]: I1206 03:40:39.640121 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2392c238-d47c-472f-b8f3-cf6ad2720454-catalog-content\") pod \"redhat-marketplace-tgld8\" (UID: \"2392c238-d47c-472f-b8f3-cf6ad2720454\") " pod="openshift-marketplace/redhat-marketplace-tgld8" Dec 06 03:40:39 crc kubenswrapper[4980]: I1206 03:40:39.640150 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bcmq5\" (UniqueName: \"kubernetes.io/projected/2392c238-d47c-472f-b8f3-cf6ad2720454-kube-api-access-bcmq5\") pod \"redhat-marketplace-tgld8\" (UID: \"2392c238-d47c-472f-b8f3-cf6ad2720454\") " pod="openshift-marketplace/redhat-marketplace-tgld8" Dec 06 03:40:39 crc kubenswrapper[4980]: I1206 03:40:39.640182 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/aca9683f-a0d4-46f8-aacb-f2e45896458c-utilities\") pod \"redhat-operators-kfc6d\" (UID: \"aca9683f-a0d4-46f8-aacb-f2e45896458c\") " pod="openshift-marketplace/redhat-operators-kfc6d" Dec 06 03:40:39 crc kubenswrapper[4980]: I1206 03:40:39.640222 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2392c238-d47c-472f-b8f3-cf6ad2720454-utilities\") pod \"redhat-marketplace-tgld8\" (UID: \"2392c238-d47c-472f-b8f3-cf6ad2720454\") " pod="openshift-marketplace/redhat-marketplace-tgld8" Dec 06 03:40:39 crc kubenswrapper[4980]: I1206 03:40:39.640253 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v4clz\" (UniqueName: \"kubernetes.io/projected/aca9683f-a0d4-46f8-aacb-f2e45896458c-kube-api-access-v4clz\") pod \"redhat-operators-kfc6d\" (UID: \"aca9683f-a0d4-46f8-aacb-f2e45896458c\") " pod="openshift-marketplace/redhat-operators-kfc6d" Dec 06 03:40:39 crc kubenswrapper[4980]: I1206 03:40:39.640996 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2392c238-d47c-472f-b8f3-cf6ad2720454-catalog-content\") pod \"redhat-marketplace-tgld8\" (UID: \"2392c238-d47c-472f-b8f3-cf6ad2720454\") " pod="openshift-marketplace/redhat-marketplace-tgld8" Dec 06 03:40:39 crc kubenswrapper[4980]: I1206 03:40:39.641230 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2392c238-d47c-472f-b8f3-cf6ad2720454-utilities\") pod \"redhat-marketplace-tgld8\" (UID: \"2392c238-d47c-472f-b8f3-cf6ad2720454\") " pod="openshift-marketplace/redhat-marketplace-tgld8" Dec 06 03:40:39 crc kubenswrapper[4980]: I1206 03:40:39.674586 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bcmq5\" (UniqueName: \"kubernetes.io/projected/2392c238-d47c-472f-b8f3-cf6ad2720454-kube-api-access-bcmq5\") pod \"redhat-marketplace-tgld8\" (UID: \"2392c238-d47c-472f-b8f3-cf6ad2720454\") " pod="openshift-marketplace/redhat-marketplace-tgld8" Dec 06 03:40:39 crc kubenswrapper[4980]: I1206 03:40:39.724852 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-tgld8" Dec 06 03:40:39 crc kubenswrapper[4980]: I1206 03:40:39.741138 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/aca9683f-a0d4-46f8-aacb-f2e45896458c-utilities\") pod \"redhat-operators-kfc6d\" (UID: \"aca9683f-a0d4-46f8-aacb-f2e45896458c\") " pod="openshift-marketplace/redhat-operators-kfc6d" Dec 06 03:40:39 crc kubenswrapper[4980]: I1206 03:40:39.741189 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v4clz\" (UniqueName: \"kubernetes.io/projected/aca9683f-a0d4-46f8-aacb-f2e45896458c-kube-api-access-v4clz\") pod \"redhat-operators-kfc6d\" (UID: \"aca9683f-a0d4-46f8-aacb-f2e45896458c\") " pod="openshift-marketplace/redhat-operators-kfc6d" Dec 06 03:40:39 crc kubenswrapper[4980]: I1206 03:40:39.741264 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/aca9683f-a0d4-46f8-aacb-f2e45896458c-catalog-content\") pod \"redhat-operators-kfc6d\" (UID: \"aca9683f-a0d4-46f8-aacb-f2e45896458c\") " pod="openshift-marketplace/redhat-operators-kfc6d" Dec 06 03:40:39 crc kubenswrapper[4980]: I1206 03:40:39.741756 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/aca9683f-a0d4-46f8-aacb-f2e45896458c-catalog-content\") pod \"redhat-operators-kfc6d\" (UID: \"aca9683f-a0d4-46f8-aacb-f2e45896458c\") " pod="openshift-marketplace/redhat-operators-kfc6d" Dec 06 03:40:39 crc kubenswrapper[4980]: I1206 03:40:39.742099 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/aca9683f-a0d4-46f8-aacb-f2e45896458c-utilities\") pod \"redhat-operators-kfc6d\" (UID: \"aca9683f-a0d4-46f8-aacb-f2e45896458c\") " pod="openshift-marketplace/redhat-operators-kfc6d" Dec 06 03:40:39 crc kubenswrapper[4980]: I1206 03:40:39.759891 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v4clz\" (UniqueName: \"kubernetes.io/projected/aca9683f-a0d4-46f8-aacb-f2e45896458c-kube-api-access-v4clz\") pod \"redhat-operators-kfc6d\" (UID: \"aca9683f-a0d4-46f8-aacb-f2e45896458c\") " pod="openshift-marketplace/redhat-operators-kfc6d" Dec 06 03:40:39 crc kubenswrapper[4980]: I1206 03:40:39.954057 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-kfc6d" Dec 06 03:40:40 crc kubenswrapper[4980]: I1206 03:40:40.028484 4980 generic.go:334] "Generic (PLEG): container finished" podID="b6205a6a-fc2b-4a1e-aefd-b86867f3478d" containerID="a599aad537a0746ae23da8065a9ba013d55d31face14f64479b864ab7f0ba76d" exitCode=0 Dec 06 03:40:40 crc kubenswrapper[4980]: I1206 03:40:40.028839 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-x8x8b" event={"ID":"b6205a6a-fc2b-4a1e-aefd-b86867f3478d","Type":"ContainerDied","Data":"a599aad537a0746ae23da8065a9ba013d55d31face14f64479b864ab7f0ba76d"} Dec 06 03:40:40 crc kubenswrapper[4980]: I1206 03:40:40.233898 4980 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-tgld8"] Dec 06 03:40:40 crc kubenswrapper[4980]: I1206 03:40:40.364469 4980 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-kfc6d"] Dec 06 03:40:40 crc kubenswrapper[4980]: W1206 03:40:40.374005 4980 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podaca9683f_a0d4_46f8_aacb_f2e45896458c.slice/crio-a07faab990630d7e5fab5d1fefb2b48c75b6e0f0654a8e50147a8ebd5ebcc40e WatchSource:0}: Error finding container a07faab990630d7e5fab5d1fefb2b48c75b6e0f0654a8e50147a8ebd5ebcc40e: Status 404 returned error can't find the container with id a07faab990630d7e5fab5d1fefb2b48c75b6e0f0654a8e50147a8ebd5ebcc40e Dec 06 03:40:41 crc kubenswrapper[4980]: I1206 03:40:41.036186 4980 generic.go:334] "Generic (PLEG): container finished" podID="aca9683f-a0d4-46f8-aacb-f2e45896458c" containerID="ef8569cb725c199a446547210646fa79ec6753797c5b8d6e317c3e2989efa064" exitCode=0 Dec 06 03:40:41 crc kubenswrapper[4980]: I1206 03:40:41.036243 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kfc6d" event={"ID":"aca9683f-a0d4-46f8-aacb-f2e45896458c","Type":"ContainerDied","Data":"ef8569cb725c199a446547210646fa79ec6753797c5b8d6e317c3e2989efa064"} Dec 06 03:40:41 crc kubenswrapper[4980]: I1206 03:40:41.036546 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kfc6d" event={"ID":"aca9683f-a0d4-46f8-aacb-f2e45896458c","Type":"ContainerStarted","Data":"a07faab990630d7e5fab5d1fefb2b48c75b6e0f0654a8e50147a8ebd5ebcc40e"} Dec 06 03:40:41 crc kubenswrapper[4980]: I1206 03:40:41.039331 4980 generic.go:334] "Generic (PLEG): container finished" podID="2392c238-d47c-472f-b8f3-cf6ad2720454" containerID="6c013a4d38d625798ba29917af1ca9e92cc284c80287bcdcf450b989cabb9a76" exitCode=0 Dec 06 03:40:41 crc kubenswrapper[4980]: I1206 03:40:41.039398 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-tgld8" event={"ID":"2392c238-d47c-472f-b8f3-cf6ad2720454","Type":"ContainerDied","Data":"6c013a4d38d625798ba29917af1ca9e92cc284c80287bcdcf450b989cabb9a76"} Dec 06 03:40:41 crc kubenswrapper[4980]: I1206 03:40:41.039426 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-tgld8" event={"ID":"2392c238-d47c-472f-b8f3-cf6ad2720454","Type":"ContainerStarted","Data":"6950879a0a90c2e6441b188dcceba6d13d95cd7313bccbc079a8f84d2a827120"} Dec 06 03:40:41 crc kubenswrapper[4980]: I1206 03:40:41.042838 4980 generic.go:334] "Generic (PLEG): container finished" podID="6b108c98-9b6c-48f5-a6a9-2db36a4e5283" containerID="b0934e763c7a15c911eb07a51d035f45bdd515e7f671c16b942cb57f0eaa979f" exitCode=0 Dec 06 03:40:41 crc kubenswrapper[4980]: I1206 03:40:41.042874 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-d2hwb" event={"ID":"6b108c98-9b6c-48f5-a6a9-2db36a4e5283","Type":"ContainerDied","Data":"b0934e763c7a15c911eb07a51d035f45bdd515e7f671c16b942cb57f0eaa979f"} Dec 06 03:40:42 crc kubenswrapper[4980]: I1206 03:40:42.050858 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-x8x8b" event={"ID":"b6205a6a-fc2b-4a1e-aefd-b86867f3478d","Type":"ContainerStarted","Data":"f09838fb93060f0c4ccae94e0ef1d45a453e7b40a2381b0ff517a45d498d2f00"} Dec 06 03:40:42 crc kubenswrapper[4980]: I1206 03:40:42.053006 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-d2hwb" event={"ID":"6b108c98-9b6c-48f5-a6a9-2db36a4e5283","Type":"ContainerStarted","Data":"efc427fac85395366b99901c3921d099b072e80e9f643b9e21698813435534ba"} Dec 06 03:40:42 crc kubenswrapper[4980]: I1206 03:40:42.066364 4980 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-x8x8b" podStartSLOduration=3.299109949 podStartE2EDuration="6.066300767s" podCreationTimestamp="2025-12-06 03:40:36 +0000 UTC" firstStartedPulling="2025-12-06 03:40:38.014601913 +0000 UTC m=+397.259353184" lastFinishedPulling="2025-12-06 03:40:40.781792731 +0000 UTC m=+400.026544002" observedRunningTime="2025-12-06 03:40:42.065502783 +0000 UTC m=+401.310254074" watchObservedRunningTime="2025-12-06 03:40:42.066300767 +0000 UTC m=+401.311052038" Dec 06 03:40:42 crc kubenswrapper[4980]: I1206 03:40:42.099773 4980 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-d2hwb" podStartSLOduration=2.516156858 podStartE2EDuration="5.099732537s" podCreationTimestamp="2025-12-06 03:40:37 +0000 UTC" firstStartedPulling="2025-12-06 03:40:39.021386759 +0000 UTC m=+398.266138030" lastFinishedPulling="2025-12-06 03:40:41.604962438 +0000 UTC m=+400.849713709" observedRunningTime="2025-12-06 03:40:42.090265334 +0000 UTC m=+401.335016605" watchObservedRunningTime="2025-12-06 03:40:42.099732537 +0000 UTC m=+401.344483798" Dec 06 03:40:43 crc kubenswrapper[4980]: I1206 03:40:43.062589 4980 generic.go:334] "Generic (PLEG): container finished" podID="2392c238-d47c-472f-b8f3-cf6ad2720454" containerID="a5d35bc420d763bdcab14c8be776c5471e5462c1d6a603e902b70d661b1d7e86" exitCode=0 Dec 06 03:40:43 crc kubenswrapper[4980]: I1206 03:40:43.062692 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-tgld8" event={"ID":"2392c238-d47c-472f-b8f3-cf6ad2720454","Type":"ContainerDied","Data":"a5d35bc420d763bdcab14c8be776c5471e5462c1d6a603e902b70d661b1d7e86"} Dec 06 03:40:43 crc kubenswrapper[4980]: I1206 03:40:43.065153 4980 generic.go:334] "Generic (PLEG): container finished" podID="aca9683f-a0d4-46f8-aacb-f2e45896458c" containerID="27828ce45dad8a75eeb43f2e1ea5c371e32799d439fe8bf70afe97b4808db965" exitCode=0 Dec 06 03:40:43 crc kubenswrapper[4980]: I1206 03:40:43.065232 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kfc6d" event={"ID":"aca9683f-a0d4-46f8-aacb-f2e45896458c","Type":"ContainerDied","Data":"27828ce45dad8a75eeb43f2e1ea5c371e32799d439fe8bf70afe97b4808db965"} Dec 06 03:40:43 crc kubenswrapper[4980]: I1206 03:40:43.904866 4980 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-66df7c8f76-gw4xb" Dec 06 03:40:44 crc kubenswrapper[4980]: I1206 03:40:44.004524 4980 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-qw2fb"] Dec 06 03:40:44 crc kubenswrapper[4980]: I1206 03:40:44.072583 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kfc6d" event={"ID":"aca9683f-a0d4-46f8-aacb-f2e45896458c","Type":"ContainerStarted","Data":"da8f0eb3917db1e433cddab6b3fe82118e0fa29adfc396cfe2a4c7a058502d34"} Dec 06 03:40:44 crc kubenswrapper[4980]: I1206 03:40:44.078341 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-tgld8" event={"ID":"2392c238-d47c-472f-b8f3-cf6ad2720454","Type":"ContainerStarted","Data":"f43aa9683e3ccfe6e4a83cd03e947b6cc17e0bc3c1e8427a208f3f13b324a72a"} Dec 06 03:40:44 crc kubenswrapper[4980]: I1206 03:40:44.105778 4980 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-kfc6d" podStartSLOduration=2.612509744 podStartE2EDuration="5.105762146s" podCreationTimestamp="2025-12-06 03:40:39 +0000 UTC" firstStartedPulling="2025-12-06 03:40:41.037349546 +0000 UTC m=+400.282100827" lastFinishedPulling="2025-12-06 03:40:43.530601958 +0000 UTC m=+402.775353229" observedRunningTime="2025-12-06 03:40:44.104623102 +0000 UTC m=+403.349374393" watchObservedRunningTime="2025-12-06 03:40:44.105762146 +0000 UTC m=+403.350513417" Dec 06 03:40:44 crc kubenswrapper[4980]: I1206 03:40:44.143915 4980 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-tgld8" podStartSLOduration=2.462360496 podStartE2EDuration="5.143892758s" podCreationTimestamp="2025-12-06 03:40:39 +0000 UTC" firstStartedPulling="2025-12-06 03:40:41.041631404 +0000 UTC m=+400.286382675" lastFinishedPulling="2025-12-06 03:40:43.723163666 +0000 UTC m=+402.967914937" observedRunningTime="2025-12-06 03:40:44.138001632 +0000 UTC m=+403.382752893" watchObservedRunningTime="2025-12-06 03:40:44.143892758 +0000 UTC m=+403.388644029" Dec 06 03:40:47 crc kubenswrapper[4980]: I1206 03:40:47.338620 4980 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-x8x8b" Dec 06 03:40:47 crc kubenswrapper[4980]: I1206 03:40:47.341406 4980 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-x8x8b" Dec 06 03:40:47 crc kubenswrapper[4980]: I1206 03:40:47.394727 4980 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-x8x8b" Dec 06 03:40:47 crc kubenswrapper[4980]: I1206 03:40:47.568174 4980 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-d2hwb" Dec 06 03:40:47 crc kubenswrapper[4980]: I1206 03:40:47.568228 4980 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-d2hwb" Dec 06 03:40:47 crc kubenswrapper[4980]: I1206 03:40:47.624114 4980 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-d2hwb" Dec 06 03:40:48 crc kubenswrapper[4980]: I1206 03:40:48.142292 4980 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-x8x8b" Dec 06 03:40:48 crc kubenswrapper[4980]: I1206 03:40:48.145671 4980 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-d2hwb" Dec 06 03:40:49 crc kubenswrapper[4980]: I1206 03:40:49.725915 4980 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-tgld8" Dec 06 03:40:49 crc kubenswrapper[4980]: I1206 03:40:49.725983 4980 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-tgld8" Dec 06 03:40:49 crc kubenswrapper[4980]: I1206 03:40:49.794362 4980 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-tgld8" Dec 06 03:40:49 crc kubenswrapper[4980]: I1206 03:40:49.955020 4980 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-kfc6d" Dec 06 03:40:49 crc kubenswrapper[4980]: I1206 03:40:49.955795 4980 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-kfc6d" Dec 06 03:40:50 crc kubenswrapper[4980]: I1206 03:40:50.003720 4980 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-kfc6d" Dec 06 03:40:50 crc kubenswrapper[4980]: I1206 03:40:50.169358 4980 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-tgld8" Dec 06 03:40:50 crc kubenswrapper[4980]: I1206 03:40:50.183247 4980 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-kfc6d" Dec 06 03:40:55 crc kubenswrapper[4980]: I1206 03:40:55.136946 4980 patch_prober.go:28] interesting pod/machine-config-daemon-r5zfb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 03:40:55 crc kubenswrapper[4980]: I1206 03:40:55.137261 4980 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-r5zfb" podUID="320f44d1-a671-4a91-b328-a8b0fdd8f23a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 03:40:55 crc kubenswrapper[4980]: I1206 03:40:55.137303 4980 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-r5zfb" Dec 06 03:40:55 crc kubenswrapper[4980]: I1206 03:40:55.137865 4980 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"fecbfcc218b13cf32c394772ff9a934bdde92dcfdcdb184849fd10fe2e85db2e"} pod="openshift-machine-config-operator/machine-config-daemon-r5zfb" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 06 03:40:55 crc kubenswrapper[4980]: I1206 03:40:55.137911 4980 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-r5zfb" podUID="320f44d1-a671-4a91-b328-a8b0fdd8f23a" containerName="machine-config-daemon" containerID="cri-o://fecbfcc218b13cf32c394772ff9a934bdde92dcfdcdb184849fd10fe2e85db2e" gracePeriod=600 Dec 06 03:40:58 crc kubenswrapper[4980]: I1206 03:40:58.158233 4980 generic.go:334] "Generic (PLEG): container finished" podID="320f44d1-a671-4a91-b328-a8b0fdd8f23a" containerID="fecbfcc218b13cf32c394772ff9a934bdde92dcfdcdb184849fd10fe2e85db2e" exitCode=0 Dec 06 03:40:58 crc kubenswrapper[4980]: I1206 03:40:58.158392 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-r5zfb" event={"ID":"320f44d1-a671-4a91-b328-a8b0fdd8f23a","Type":"ContainerDied","Data":"fecbfcc218b13cf32c394772ff9a934bdde92dcfdcdb184849fd10fe2e85db2e"} Dec 06 03:40:58 crc kubenswrapper[4980]: I1206 03:40:58.158815 4980 scope.go:117] "RemoveContainer" containerID="52e8db644463794041978ced2c0591c13e0ada8e5f1a78a94a0a665218a37b5f" Dec 06 03:40:59 crc kubenswrapper[4980]: I1206 03:40:59.167981 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-r5zfb" event={"ID":"320f44d1-a671-4a91-b328-a8b0fdd8f23a","Type":"ContainerStarted","Data":"ca8484c03e9f51ca249527d107dc686e5fd267b68f8d80ac4706ca616b6cf584"} Dec 06 03:41:09 crc kubenswrapper[4980]: I1206 03:41:09.046999 4980 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-image-registry/image-registry-697d97f7c8-qw2fb" podUID="b25ba819-6e4e-4ca0-a570-c2726b8fbec5" containerName="registry" containerID="cri-o://9e89a295b67d24815fed2d32d770bfc36b4d92b99ab27d36045774ea8cf6e846" gracePeriod=30 Dec 06 03:41:09 crc kubenswrapper[4980]: I1206 03:41:09.276154 4980 generic.go:334] "Generic (PLEG): container finished" podID="b25ba819-6e4e-4ca0-a570-c2726b8fbec5" containerID="9e89a295b67d24815fed2d32d770bfc36b4d92b99ab27d36045774ea8cf6e846" exitCode=0 Dec 06 03:41:09 crc kubenswrapper[4980]: I1206 03:41:09.276197 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-qw2fb" event={"ID":"b25ba819-6e4e-4ca0-a570-c2726b8fbec5","Type":"ContainerDied","Data":"9e89a295b67d24815fed2d32d770bfc36b4d92b99ab27d36045774ea8cf6e846"} Dec 06 03:41:09 crc kubenswrapper[4980]: I1206 03:41:09.404800 4980 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-qw2fb" Dec 06 03:41:09 crc kubenswrapper[4980]: I1206 03:41:09.550426 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/b25ba819-6e4e-4ca0-a570-c2726b8fbec5-ca-trust-extracted\") pod \"b25ba819-6e4e-4ca0-a570-c2726b8fbec5\" (UID: \"b25ba819-6e4e-4ca0-a570-c2726b8fbec5\") " Dec 06 03:41:09 crc kubenswrapper[4980]: I1206 03:41:09.550626 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/b25ba819-6e4e-4ca0-a570-c2726b8fbec5-registry-tls\") pod \"b25ba819-6e4e-4ca0-a570-c2726b8fbec5\" (UID: \"b25ba819-6e4e-4ca0-a570-c2726b8fbec5\") " Dec 06 03:41:09 crc kubenswrapper[4980]: I1206 03:41:09.550724 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/b25ba819-6e4e-4ca0-a570-c2726b8fbec5-registry-certificates\") pod \"b25ba819-6e4e-4ca0-a570-c2726b8fbec5\" (UID: \"b25ba819-6e4e-4ca0-a570-c2726b8fbec5\") " Dec 06 03:41:09 crc kubenswrapper[4980]: I1206 03:41:09.550775 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b25ba819-6e4e-4ca0-a570-c2726b8fbec5-trusted-ca\") pod \"b25ba819-6e4e-4ca0-a570-c2726b8fbec5\" (UID: \"b25ba819-6e4e-4ca0-a570-c2726b8fbec5\") " Dec 06 03:41:09 crc kubenswrapper[4980]: I1206 03:41:09.550812 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/b25ba819-6e4e-4ca0-a570-c2726b8fbec5-installation-pull-secrets\") pod \"b25ba819-6e4e-4ca0-a570-c2726b8fbec5\" (UID: \"b25ba819-6e4e-4ca0-a570-c2726b8fbec5\") " Dec 06 03:41:09 crc kubenswrapper[4980]: I1206 03:41:09.550864 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bzls8\" (UniqueName: \"kubernetes.io/projected/b25ba819-6e4e-4ca0-a570-c2726b8fbec5-kube-api-access-bzls8\") pod \"b25ba819-6e4e-4ca0-a570-c2726b8fbec5\" (UID: \"b25ba819-6e4e-4ca0-a570-c2726b8fbec5\") " Dec 06 03:41:09 crc kubenswrapper[4980]: I1206 03:41:09.551207 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-storage\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"b25ba819-6e4e-4ca0-a570-c2726b8fbec5\" (UID: \"b25ba819-6e4e-4ca0-a570-c2726b8fbec5\") " Dec 06 03:41:09 crc kubenswrapper[4980]: I1206 03:41:09.551272 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/b25ba819-6e4e-4ca0-a570-c2726b8fbec5-bound-sa-token\") pod \"b25ba819-6e4e-4ca0-a570-c2726b8fbec5\" (UID: \"b25ba819-6e4e-4ca0-a570-c2726b8fbec5\") " Dec 06 03:41:09 crc kubenswrapper[4980]: I1206 03:41:09.551694 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b25ba819-6e4e-4ca0-a570-c2726b8fbec5-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "b25ba819-6e4e-4ca0-a570-c2726b8fbec5" (UID: "b25ba819-6e4e-4ca0-a570-c2726b8fbec5"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 03:41:09 crc kubenswrapper[4980]: I1206 03:41:09.552197 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b25ba819-6e4e-4ca0-a570-c2726b8fbec5-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "b25ba819-6e4e-4ca0-a570-c2726b8fbec5" (UID: "b25ba819-6e4e-4ca0-a570-c2726b8fbec5"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 03:41:09 crc kubenswrapper[4980]: I1206 03:41:09.557606 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b25ba819-6e4e-4ca0-a570-c2726b8fbec5-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "b25ba819-6e4e-4ca0-a570-c2726b8fbec5" (UID: "b25ba819-6e4e-4ca0-a570-c2726b8fbec5"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 03:41:09 crc kubenswrapper[4980]: I1206 03:41:09.557623 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b25ba819-6e4e-4ca0-a570-c2726b8fbec5-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "b25ba819-6e4e-4ca0-a570-c2726b8fbec5" (UID: "b25ba819-6e4e-4ca0-a570-c2726b8fbec5"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 03:41:09 crc kubenswrapper[4980]: I1206 03:41:09.560806 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b25ba819-6e4e-4ca0-a570-c2726b8fbec5-kube-api-access-bzls8" (OuterVolumeSpecName: "kube-api-access-bzls8") pod "b25ba819-6e4e-4ca0-a570-c2726b8fbec5" (UID: "b25ba819-6e4e-4ca0-a570-c2726b8fbec5"). InnerVolumeSpecName "kube-api-access-bzls8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 03:41:09 crc kubenswrapper[4980]: I1206 03:41:09.561411 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "registry-storage") pod "b25ba819-6e4e-4ca0-a570-c2726b8fbec5" (UID: "b25ba819-6e4e-4ca0-a570-c2726b8fbec5"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Dec 06 03:41:09 crc kubenswrapper[4980]: I1206 03:41:09.563083 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b25ba819-6e4e-4ca0-a570-c2726b8fbec5-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "b25ba819-6e4e-4ca0-a570-c2726b8fbec5" (UID: "b25ba819-6e4e-4ca0-a570-c2726b8fbec5"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 03:41:09 crc kubenswrapper[4980]: I1206 03:41:09.577669 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b25ba819-6e4e-4ca0-a570-c2726b8fbec5-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "b25ba819-6e4e-4ca0-a570-c2726b8fbec5" (UID: "b25ba819-6e4e-4ca0-a570-c2726b8fbec5"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 03:41:09 crc kubenswrapper[4980]: I1206 03:41:09.652679 4980 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/b25ba819-6e4e-4ca0-a570-c2726b8fbec5-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Dec 06 03:41:09 crc kubenswrapper[4980]: I1206 03:41:09.652728 4980 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/b25ba819-6e4e-4ca0-a570-c2726b8fbec5-registry-tls\") on node \"crc\" DevicePath \"\"" Dec 06 03:41:09 crc kubenswrapper[4980]: I1206 03:41:09.652741 4980 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/b25ba819-6e4e-4ca0-a570-c2726b8fbec5-registry-certificates\") on node \"crc\" DevicePath \"\"" Dec 06 03:41:09 crc kubenswrapper[4980]: I1206 03:41:09.652755 4980 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b25ba819-6e4e-4ca0-a570-c2726b8fbec5-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 06 03:41:09 crc kubenswrapper[4980]: I1206 03:41:09.652767 4980 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/b25ba819-6e4e-4ca0-a570-c2726b8fbec5-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Dec 06 03:41:09 crc kubenswrapper[4980]: I1206 03:41:09.652779 4980 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bzls8\" (UniqueName: \"kubernetes.io/projected/b25ba819-6e4e-4ca0-a570-c2726b8fbec5-kube-api-access-bzls8\") on node \"crc\" DevicePath \"\"" Dec 06 03:41:09 crc kubenswrapper[4980]: I1206 03:41:09.652790 4980 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/b25ba819-6e4e-4ca0-a570-c2726b8fbec5-bound-sa-token\") on node \"crc\" DevicePath \"\"" Dec 06 03:41:10 crc kubenswrapper[4980]: I1206 03:41:10.283494 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-qw2fb" event={"ID":"b25ba819-6e4e-4ca0-a570-c2726b8fbec5","Type":"ContainerDied","Data":"6f1a12c733903ff72d2e17f9c24273a497ba9798f9450b0680a991fe69b18a14"} Dec 06 03:41:10 crc kubenswrapper[4980]: I1206 03:41:10.283595 4980 scope.go:117] "RemoveContainer" containerID="9e89a295b67d24815fed2d32d770bfc36b4d92b99ab27d36045774ea8cf6e846" Dec 06 03:41:10 crc kubenswrapper[4980]: I1206 03:41:10.283625 4980 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-qw2fb" Dec 06 03:41:10 crc kubenswrapper[4980]: I1206 03:41:10.328635 4980 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-qw2fb"] Dec 06 03:41:10 crc kubenswrapper[4980]: I1206 03:41:10.335226 4980 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-qw2fb"] Dec 06 03:41:11 crc kubenswrapper[4980]: I1206 03:41:11.197781 4980 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b25ba819-6e4e-4ca0-a570-c2726b8fbec5" path="/var/lib/kubelet/pods/b25ba819-6e4e-4ca0-a570-c2726b8fbec5/volumes" Dec 06 03:43:25 crc kubenswrapper[4980]: I1206 03:43:25.137828 4980 patch_prober.go:28] interesting pod/machine-config-daemon-r5zfb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 03:43:25 crc kubenswrapper[4980]: I1206 03:43:25.138453 4980 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-r5zfb" podUID="320f44d1-a671-4a91-b328-a8b0fdd8f23a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 03:43:55 crc kubenswrapper[4980]: I1206 03:43:55.137353 4980 patch_prober.go:28] interesting pod/machine-config-daemon-r5zfb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 03:43:55 crc kubenswrapper[4980]: I1206 03:43:55.138159 4980 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-r5zfb" podUID="320f44d1-a671-4a91-b328-a8b0fdd8f23a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 03:44:25 crc kubenswrapper[4980]: I1206 03:44:25.137813 4980 patch_prober.go:28] interesting pod/machine-config-daemon-r5zfb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 03:44:25 crc kubenswrapper[4980]: I1206 03:44:25.138436 4980 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-r5zfb" podUID="320f44d1-a671-4a91-b328-a8b0fdd8f23a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 03:44:25 crc kubenswrapper[4980]: I1206 03:44:25.138562 4980 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-r5zfb" Dec 06 03:44:25 crc kubenswrapper[4980]: I1206 03:44:25.139234 4980 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"ca8484c03e9f51ca249527d107dc686e5fd267b68f8d80ac4706ca616b6cf584"} pod="openshift-machine-config-operator/machine-config-daemon-r5zfb" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 06 03:44:25 crc kubenswrapper[4980]: I1206 03:44:25.139370 4980 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-r5zfb" podUID="320f44d1-a671-4a91-b328-a8b0fdd8f23a" containerName="machine-config-daemon" containerID="cri-o://ca8484c03e9f51ca249527d107dc686e5fd267b68f8d80ac4706ca616b6cf584" gracePeriod=600 Dec 06 03:44:25 crc kubenswrapper[4980]: I1206 03:44:25.851453 4980 generic.go:334] "Generic (PLEG): container finished" podID="320f44d1-a671-4a91-b328-a8b0fdd8f23a" containerID="ca8484c03e9f51ca249527d107dc686e5fd267b68f8d80ac4706ca616b6cf584" exitCode=0 Dec 06 03:44:25 crc kubenswrapper[4980]: I1206 03:44:25.851583 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-r5zfb" event={"ID":"320f44d1-a671-4a91-b328-a8b0fdd8f23a","Type":"ContainerDied","Data":"ca8484c03e9f51ca249527d107dc686e5fd267b68f8d80ac4706ca616b6cf584"} Dec 06 03:44:25 crc kubenswrapper[4980]: I1206 03:44:25.851839 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-r5zfb" event={"ID":"320f44d1-a671-4a91-b328-a8b0fdd8f23a","Type":"ContainerStarted","Data":"c1b55d544de848b8b12ac6ab1d4d7ce58e20980d22589239fd7500d5b466d62e"} Dec 06 03:44:25 crc kubenswrapper[4980]: I1206 03:44:25.851884 4980 scope.go:117] "RemoveContainer" containerID="fecbfcc218b13cf32c394772ff9a934bdde92dcfdcdb184849fd10fe2e85db2e" Dec 06 03:45:00 crc kubenswrapper[4980]: I1206 03:45:00.188687 4980 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29416545-d8hns"] Dec 06 03:45:00 crc kubenswrapper[4980]: E1206 03:45:00.190580 4980 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b25ba819-6e4e-4ca0-a570-c2726b8fbec5" containerName="registry" Dec 06 03:45:00 crc kubenswrapper[4980]: I1206 03:45:00.190678 4980 state_mem.go:107] "Deleted CPUSet assignment" podUID="b25ba819-6e4e-4ca0-a570-c2726b8fbec5" containerName="registry" Dec 06 03:45:00 crc kubenswrapper[4980]: I1206 03:45:00.190875 4980 memory_manager.go:354] "RemoveStaleState removing state" podUID="b25ba819-6e4e-4ca0-a570-c2726b8fbec5" containerName="registry" Dec 06 03:45:00 crc kubenswrapper[4980]: I1206 03:45:00.191575 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29416545-d8hns" Dec 06 03:45:00 crc kubenswrapper[4980]: I1206 03:45:00.193650 4980 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 06 03:45:00 crc kubenswrapper[4980]: I1206 03:45:00.193664 4980 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 06 03:45:00 crc kubenswrapper[4980]: I1206 03:45:00.203247 4980 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29416545-d8hns"] Dec 06 03:45:00 crc kubenswrapper[4980]: I1206 03:45:00.326211 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/30114d1d-a838-4c2c-9029-5d3faa13de1d-secret-volume\") pod \"collect-profiles-29416545-d8hns\" (UID: \"30114d1d-a838-4c2c-9029-5d3faa13de1d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416545-d8hns" Dec 06 03:45:00 crc kubenswrapper[4980]: I1206 03:45:00.326336 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qdv2x\" (UniqueName: \"kubernetes.io/projected/30114d1d-a838-4c2c-9029-5d3faa13de1d-kube-api-access-qdv2x\") pod \"collect-profiles-29416545-d8hns\" (UID: \"30114d1d-a838-4c2c-9029-5d3faa13de1d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416545-d8hns" Dec 06 03:45:00 crc kubenswrapper[4980]: I1206 03:45:00.326391 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/30114d1d-a838-4c2c-9029-5d3faa13de1d-config-volume\") pod \"collect-profiles-29416545-d8hns\" (UID: \"30114d1d-a838-4c2c-9029-5d3faa13de1d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416545-d8hns" Dec 06 03:45:00 crc kubenswrapper[4980]: I1206 03:45:00.427688 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/30114d1d-a838-4c2c-9029-5d3faa13de1d-config-volume\") pod \"collect-profiles-29416545-d8hns\" (UID: \"30114d1d-a838-4c2c-9029-5d3faa13de1d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416545-d8hns" Dec 06 03:45:00 crc kubenswrapper[4980]: I1206 03:45:00.427816 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/30114d1d-a838-4c2c-9029-5d3faa13de1d-secret-volume\") pod \"collect-profiles-29416545-d8hns\" (UID: \"30114d1d-a838-4c2c-9029-5d3faa13de1d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416545-d8hns" Dec 06 03:45:00 crc kubenswrapper[4980]: I1206 03:45:00.427871 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qdv2x\" (UniqueName: \"kubernetes.io/projected/30114d1d-a838-4c2c-9029-5d3faa13de1d-kube-api-access-qdv2x\") pod \"collect-profiles-29416545-d8hns\" (UID: \"30114d1d-a838-4c2c-9029-5d3faa13de1d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416545-d8hns" Dec 06 03:45:00 crc kubenswrapper[4980]: I1206 03:45:00.431015 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/30114d1d-a838-4c2c-9029-5d3faa13de1d-config-volume\") pod \"collect-profiles-29416545-d8hns\" (UID: \"30114d1d-a838-4c2c-9029-5d3faa13de1d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416545-d8hns" Dec 06 03:45:00 crc kubenswrapper[4980]: I1206 03:45:00.440258 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/30114d1d-a838-4c2c-9029-5d3faa13de1d-secret-volume\") pod \"collect-profiles-29416545-d8hns\" (UID: \"30114d1d-a838-4c2c-9029-5d3faa13de1d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416545-d8hns" Dec 06 03:45:00 crc kubenswrapper[4980]: I1206 03:45:00.448828 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qdv2x\" (UniqueName: \"kubernetes.io/projected/30114d1d-a838-4c2c-9029-5d3faa13de1d-kube-api-access-qdv2x\") pod \"collect-profiles-29416545-d8hns\" (UID: \"30114d1d-a838-4c2c-9029-5d3faa13de1d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416545-d8hns" Dec 06 03:45:00 crc kubenswrapper[4980]: I1206 03:45:00.511555 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29416545-d8hns" Dec 06 03:45:00 crc kubenswrapper[4980]: I1206 03:45:00.702737 4980 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29416545-d8hns"] Dec 06 03:45:01 crc kubenswrapper[4980]: I1206 03:45:01.100108 4980 generic.go:334] "Generic (PLEG): container finished" podID="30114d1d-a838-4c2c-9029-5d3faa13de1d" containerID="bc210483d41a1123e98c24ee5698650fa3ef3f4bf88a454c7228215c73543dec" exitCode=0 Dec 06 03:45:01 crc kubenswrapper[4980]: I1206 03:45:01.100177 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29416545-d8hns" event={"ID":"30114d1d-a838-4c2c-9029-5d3faa13de1d","Type":"ContainerDied","Data":"bc210483d41a1123e98c24ee5698650fa3ef3f4bf88a454c7228215c73543dec"} Dec 06 03:45:01 crc kubenswrapper[4980]: I1206 03:45:01.100245 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29416545-d8hns" event={"ID":"30114d1d-a838-4c2c-9029-5d3faa13de1d","Type":"ContainerStarted","Data":"53e6003b3094b837e4d018b63459cf4ae9458f725e669dc1101e2768d619da8b"} Dec 06 03:45:02 crc kubenswrapper[4980]: I1206 03:45:02.408428 4980 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29416545-d8hns" Dec 06 03:45:02 crc kubenswrapper[4980]: I1206 03:45:02.455478 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/30114d1d-a838-4c2c-9029-5d3faa13de1d-secret-volume\") pod \"30114d1d-a838-4c2c-9029-5d3faa13de1d\" (UID: \"30114d1d-a838-4c2c-9029-5d3faa13de1d\") " Dec 06 03:45:02 crc kubenswrapper[4980]: I1206 03:45:02.455554 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qdv2x\" (UniqueName: \"kubernetes.io/projected/30114d1d-a838-4c2c-9029-5d3faa13de1d-kube-api-access-qdv2x\") pod \"30114d1d-a838-4c2c-9029-5d3faa13de1d\" (UID: \"30114d1d-a838-4c2c-9029-5d3faa13de1d\") " Dec 06 03:45:02 crc kubenswrapper[4980]: I1206 03:45:02.455603 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/30114d1d-a838-4c2c-9029-5d3faa13de1d-config-volume\") pod \"30114d1d-a838-4c2c-9029-5d3faa13de1d\" (UID: \"30114d1d-a838-4c2c-9029-5d3faa13de1d\") " Dec 06 03:45:02 crc kubenswrapper[4980]: I1206 03:45:02.456321 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/30114d1d-a838-4c2c-9029-5d3faa13de1d-config-volume" (OuterVolumeSpecName: "config-volume") pod "30114d1d-a838-4c2c-9029-5d3faa13de1d" (UID: "30114d1d-a838-4c2c-9029-5d3faa13de1d"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 03:45:02 crc kubenswrapper[4980]: I1206 03:45:02.460580 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/30114d1d-a838-4c2c-9029-5d3faa13de1d-kube-api-access-qdv2x" (OuterVolumeSpecName: "kube-api-access-qdv2x") pod "30114d1d-a838-4c2c-9029-5d3faa13de1d" (UID: "30114d1d-a838-4c2c-9029-5d3faa13de1d"). InnerVolumeSpecName "kube-api-access-qdv2x". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 03:45:02 crc kubenswrapper[4980]: I1206 03:45:02.460624 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/30114d1d-a838-4c2c-9029-5d3faa13de1d-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "30114d1d-a838-4c2c-9029-5d3faa13de1d" (UID: "30114d1d-a838-4c2c-9029-5d3faa13de1d"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 03:45:02 crc kubenswrapper[4980]: I1206 03:45:02.557638 4980 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/30114d1d-a838-4c2c-9029-5d3faa13de1d-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 06 03:45:02 crc kubenswrapper[4980]: I1206 03:45:02.557715 4980 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qdv2x\" (UniqueName: \"kubernetes.io/projected/30114d1d-a838-4c2c-9029-5d3faa13de1d-kube-api-access-qdv2x\") on node \"crc\" DevicePath \"\"" Dec 06 03:45:02 crc kubenswrapper[4980]: I1206 03:45:02.557746 4980 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/30114d1d-a838-4c2c-9029-5d3faa13de1d-config-volume\") on node \"crc\" DevicePath \"\"" Dec 06 03:45:03 crc kubenswrapper[4980]: I1206 03:45:03.116133 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29416545-d8hns" event={"ID":"30114d1d-a838-4c2c-9029-5d3faa13de1d","Type":"ContainerDied","Data":"53e6003b3094b837e4d018b63459cf4ae9458f725e669dc1101e2768d619da8b"} Dec 06 03:45:03 crc kubenswrapper[4980]: I1206 03:45:03.116485 4980 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="53e6003b3094b837e4d018b63459cf4ae9458f725e669dc1101e2768d619da8b" Dec 06 03:45:03 crc kubenswrapper[4980]: I1206 03:45:03.116184 4980 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29416545-d8hns" Dec 06 03:46:15 crc kubenswrapper[4980]: I1206 03:46:15.741630 4980 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-5w4fr"] Dec 06 03:46:15 crc kubenswrapper[4980]: I1206 03:46:15.742677 4980 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-5w4fr" podUID="5af0e768-3c9c-4401-ab56-57516cd3170e" containerName="ovn-controller" containerID="cri-o://a940ec6ee523313728f7e8ae934b7e0b0cff82771e3e64723309db1fae013c3a" gracePeriod=30 Dec 06 03:46:15 crc kubenswrapper[4980]: I1206 03:46:15.742720 4980 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-5w4fr" podUID="5af0e768-3c9c-4401-ab56-57516cd3170e" containerName="kube-rbac-proxy-ovn-metrics" containerID="cri-o://6342903b318317ee8d84c2f5d467cb25c811a789be0d6ab0e0cba37b5611e16d" gracePeriod=30 Dec 06 03:46:15 crc kubenswrapper[4980]: I1206 03:46:15.742722 4980 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-5w4fr" podUID="5af0e768-3c9c-4401-ab56-57516cd3170e" containerName="nbdb" containerID="cri-o://2d620685574db7594132ae24a91ec0a93dce20a9a289540deebb39650b8a1816" gracePeriod=30 Dec 06 03:46:15 crc kubenswrapper[4980]: I1206 03:46:15.742816 4980 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-5w4fr" podUID="5af0e768-3c9c-4401-ab56-57516cd3170e" containerName="northd" containerID="cri-o://c67ec8f9c1764fcdf215d52822c2d4f0437dc7f056f05d708767121f6c1f3515" gracePeriod=30 Dec 06 03:46:15 crc kubenswrapper[4980]: I1206 03:46:15.742825 4980 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-5w4fr" podUID="5af0e768-3c9c-4401-ab56-57516cd3170e" containerName="kube-rbac-proxy-node" containerID="cri-o://768b8e1fe1a00a8467d744da5ad3f61006b0a3612c96f8e4d0ee6fcb93195f9f" gracePeriod=30 Dec 06 03:46:15 crc kubenswrapper[4980]: I1206 03:46:15.742871 4980 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-5w4fr" podUID="5af0e768-3c9c-4401-ab56-57516cd3170e" containerName="ovn-acl-logging" containerID="cri-o://f35beae5339eeaf26494ed678e3c1bd02349eae2a4b4615325b17923f3b653b9" gracePeriod=30 Dec 06 03:46:15 crc kubenswrapper[4980]: I1206 03:46:15.742967 4980 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-5w4fr" podUID="5af0e768-3c9c-4401-ab56-57516cd3170e" containerName="sbdb" containerID="cri-o://fe261f4322b85702181d7473ab2a9a4d2d4b869d586c3091f811b12a81a63aae" gracePeriod=30 Dec 06 03:46:15 crc kubenswrapper[4980]: I1206 03:46:15.777447 4980 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-5w4fr" podUID="5af0e768-3c9c-4401-ab56-57516cd3170e" containerName="ovnkube-controller" containerID="cri-o://2d3ca7fa81b93eca00b5d90c8b2685ff607a8156fe318cf99c5296fd65a5be75" gracePeriod=30 Dec 06 03:46:16 crc kubenswrapper[4980]: I1206 03:46:16.077251 4980 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-5w4fr_5af0e768-3c9c-4401-ab56-57516cd3170e/ovnkube-controller/3.log" Dec 06 03:46:16 crc kubenswrapper[4980]: I1206 03:46:16.079432 4980 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-5w4fr_5af0e768-3c9c-4401-ab56-57516cd3170e/ovn-acl-logging/0.log" Dec 06 03:46:16 crc kubenswrapper[4980]: I1206 03:46:16.079906 4980 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-5w4fr_5af0e768-3c9c-4401-ab56-57516cd3170e/ovn-controller/0.log" Dec 06 03:46:16 crc kubenswrapper[4980]: I1206 03:46:16.080280 4980 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-5w4fr" Dec 06 03:46:16 crc kubenswrapper[4980]: I1206 03:46:16.137786 4980 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-6hnct"] Dec 06 03:46:16 crc kubenswrapper[4980]: E1206 03:46:16.138077 4980 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5af0e768-3c9c-4401-ab56-57516cd3170e" containerName="ovn-acl-logging" Dec 06 03:46:16 crc kubenswrapper[4980]: I1206 03:46:16.138096 4980 state_mem.go:107] "Deleted CPUSet assignment" podUID="5af0e768-3c9c-4401-ab56-57516cd3170e" containerName="ovn-acl-logging" Dec 06 03:46:16 crc kubenswrapper[4980]: E1206 03:46:16.138118 4980 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5af0e768-3c9c-4401-ab56-57516cd3170e" containerName="ovnkube-controller" Dec 06 03:46:16 crc kubenswrapper[4980]: I1206 03:46:16.138128 4980 state_mem.go:107] "Deleted CPUSet assignment" podUID="5af0e768-3c9c-4401-ab56-57516cd3170e" containerName="ovnkube-controller" Dec 06 03:46:16 crc kubenswrapper[4980]: E1206 03:46:16.138141 4980 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5af0e768-3c9c-4401-ab56-57516cd3170e" containerName="northd" Dec 06 03:46:16 crc kubenswrapper[4980]: I1206 03:46:16.138149 4980 state_mem.go:107] "Deleted CPUSet assignment" podUID="5af0e768-3c9c-4401-ab56-57516cd3170e" containerName="northd" Dec 06 03:46:16 crc kubenswrapper[4980]: E1206 03:46:16.138164 4980 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5af0e768-3c9c-4401-ab56-57516cd3170e" containerName="kubecfg-setup" Dec 06 03:46:16 crc kubenswrapper[4980]: I1206 03:46:16.138172 4980 state_mem.go:107] "Deleted CPUSet assignment" podUID="5af0e768-3c9c-4401-ab56-57516cd3170e" containerName="kubecfg-setup" Dec 06 03:46:16 crc kubenswrapper[4980]: E1206 03:46:16.138182 4980 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5af0e768-3c9c-4401-ab56-57516cd3170e" containerName="ovnkube-controller" Dec 06 03:46:16 crc kubenswrapper[4980]: I1206 03:46:16.138190 4980 state_mem.go:107] "Deleted CPUSet assignment" podUID="5af0e768-3c9c-4401-ab56-57516cd3170e" containerName="ovnkube-controller" Dec 06 03:46:16 crc kubenswrapper[4980]: E1206 03:46:16.138201 4980 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5af0e768-3c9c-4401-ab56-57516cd3170e" containerName="sbdb" Dec 06 03:46:16 crc kubenswrapper[4980]: I1206 03:46:16.138208 4980 state_mem.go:107] "Deleted CPUSet assignment" podUID="5af0e768-3c9c-4401-ab56-57516cd3170e" containerName="sbdb" Dec 06 03:46:16 crc kubenswrapper[4980]: E1206 03:46:16.138217 4980 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="30114d1d-a838-4c2c-9029-5d3faa13de1d" containerName="collect-profiles" Dec 06 03:46:16 crc kubenswrapper[4980]: I1206 03:46:16.138225 4980 state_mem.go:107] "Deleted CPUSet assignment" podUID="30114d1d-a838-4c2c-9029-5d3faa13de1d" containerName="collect-profiles" Dec 06 03:46:16 crc kubenswrapper[4980]: E1206 03:46:16.138234 4980 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5af0e768-3c9c-4401-ab56-57516cd3170e" containerName="ovnkube-controller" Dec 06 03:46:16 crc kubenswrapper[4980]: I1206 03:46:16.138242 4980 state_mem.go:107] "Deleted CPUSet assignment" podUID="5af0e768-3c9c-4401-ab56-57516cd3170e" containerName="ovnkube-controller" Dec 06 03:46:16 crc kubenswrapper[4980]: E1206 03:46:16.138250 4980 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5af0e768-3c9c-4401-ab56-57516cd3170e" containerName="kube-rbac-proxy-ovn-metrics" Dec 06 03:46:16 crc kubenswrapper[4980]: I1206 03:46:16.138258 4980 state_mem.go:107] "Deleted CPUSet assignment" podUID="5af0e768-3c9c-4401-ab56-57516cd3170e" containerName="kube-rbac-proxy-ovn-metrics" Dec 06 03:46:16 crc kubenswrapper[4980]: E1206 03:46:16.138269 4980 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5af0e768-3c9c-4401-ab56-57516cd3170e" containerName="ovnkube-controller" Dec 06 03:46:16 crc kubenswrapper[4980]: I1206 03:46:16.138277 4980 state_mem.go:107] "Deleted CPUSet assignment" podUID="5af0e768-3c9c-4401-ab56-57516cd3170e" containerName="ovnkube-controller" Dec 06 03:46:16 crc kubenswrapper[4980]: E1206 03:46:16.138289 4980 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5af0e768-3c9c-4401-ab56-57516cd3170e" containerName="ovn-controller" Dec 06 03:46:16 crc kubenswrapper[4980]: I1206 03:46:16.138297 4980 state_mem.go:107] "Deleted CPUSet assignment" podUID="5af0e768-3c9c-4401-ab56-57516cd3170e" containerName="ovn-controller" Dec 06 03:46:16 crc kubenswrapper[4980]: E1206 03:46:16.138311 4980 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5af0e768-3c9c-4401-ab56-57516cd3170e" containerName="kube-rbac-proxy-node" Dec 06 03:46:16 crc kubenswrapper[4980]: I1206 03:46:16.138319 4980 state_mem.go:107] "Deleted CPUSet assignment" podUID="5af0e768-3c9c-4401-ab56-57516cd3170e" containerName="kube-rbac-proxy-node" Dec 06 03:46:16 crc kubenswrapper[4980]: E1206 03:46:16.138329 4980 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5af0e768-3c9c-4401-ab56-57516cd3170e" containerName="nbdb" Dec 06 03:46:16 crc kubenswrapper[4980]: I1206 03:46:16.138337 4980 state_mem.go:107] "Deleted CPUSet assignment" podUID="5af0e768-3c9c-4401-ab56-57516cd3170e" containerName="nbdb" Dec 06 03:46:16 crc kubenswrapper[4980]: I1206 03:46:16.138445 4980 memory_manager.go:354] "RemoveStaleState removing state" podUID="5af0e768-3c9c-4401-ab56-57516cd3170e" containerName="ovnkube-controller" Dec 06 03:46:16 crc kubenswrapper[4980]: I1206 03:46:16.138460 4980 memory_manager.go:354] "RemoveStaleState removing state" podUID="5af0e768-3c9c-4401-ab56-57516cd3170e" containerName="kube-rbac-proxy-node" Dec 06 03:46:16 crc kubenswrapper[4980]: I1206 03:46:16.138473 4980 memory_manager.go:354] "RemoveStaleState removing state" podUID="5af0e768-3c9c-4401-ab56-57516cd3170e" containerName="kube-rbac-proxy-ovn-metrics" Dec 06 03:46:16 crc kubenswrapper[4980]: I1206 03:46:16.138483 4980 memory_manager.go:354] "RemoveStaleState removing state" podUID="5af0e768-3c9c-4401-ab56-57516cd3170e" containerName="ovnkube-controller" Dec 06 03:46:16 crc kubenswrapper[4980]: I1206 03:46:16.138495 4980 memory_manager.go:354] "RemoveStaleState removing state" podUID="5af0e768-3c9c-4401-ab56-57516cd3170e" containerName="ovn-acl-logging" Dec 06 03:46:16 crc kubenswrapper[4980]: I1206 03:46:16.138505 4980 memory_manager.go:354] "RemoveStaleState removing state" podUID="5af0e768-3c9c-4401-ab56-57516cd3170e" containerName="sbdb" Dec 06 03:46:16 crc kubenswrapper[4980]: I1206 03:46:16.138537 4980 memory_manager.go:354] "RemoveStaleState removing state" podUID="5af0e768-3c9c-4401-ab56-57516cd3170e" containerName="ovnkube-controller" Dec 06 03:46:16 crc kubenswrapper[4980]: I1206 03:46:16.138549 4980 memory_manager.go:354] "RemoveStaleState removing state" podUID="30114d1d-a838-4c2c-9029-5d3faa13de1d" containerName="collect-profiles" Dec 06 03:46:16 crc kubenswrapper[4980]: I1206 03:46:16.138560 4980 memory_manager.go:354] "RemoveStaleState removing state" podUID="5af0e768-3c9c-4401-ab56-57516cd3170e" containerName="northd" Dec 06 03:46:16 crc kubenswrapper[4980]: I1206 03:46:16.138572 4980 memory_manager.go:354] "RemoveStaleState removing state" podUID="5af0e768-3c9c-4401-ab56-57516cd3170e" containerName="nbdb" Dec 06 03:46:16 crc kubenswrapper[4980]: I1206 03:46:16.138584 4980 memory_manager.go:354] "RemoveStaleState removing state" podUID="5af0e768-3c9c-4401-ab56-57516cd3170e" containerName="ovn-controller" Dec 06 03:46:16 crc kubenswrapper[4980]: E1206 03:46:16.138703 4980 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5af0e768-3c9c-4401-ab56-57516cd3170e" containerName="ovnkube-controller" Dec 06 03:46:16 crc kubenswrapper[4980]: I1206 03:46:16.138714 4980 state_mem.go:107] "Deleted CPUSet assignment" podUID="5af0e768-3c9c-4401-ab56-57516cd3170e" containerName="ovnkube-controller" Dec 06 03:46:16 crc kubenswrapper[4980]: I1206 03:46:16.138828 4980 memory_manager.go:354] "RemoveStaleState removing state" podUID="5af0e768-3c9c-4401-ab56-57516cd3170e" containerName="ovnkube-controller" Dec 06 03:46:16 crc kubenswrapper[4980]: I1206 03:46:16.139058 4980 memory_manager.go:354] "RemoveStaleState removing state" podUID="5af0e768-3c9c-4401-ab56-57516cd3170e" containerName="ovnkube-controller" Dec 06 03:46:16 crc kubenswrapper[4980]: I1206 03:46:16.141178 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-6hnct" Dec 06 03:46:16 crc kubenswrapper[4980]: I1206 03:46:16.245710 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/5af0e768-3c9c-4401-ab56-57516cd3170e-log-socket\") pod \"5af0e768-3c9c-4401-ab56-57516cd3170e\" (UID: \"5af0e768-3c9c-4401-ab56-57516cd3170e\") " Dec 06 03:46:16 crc kubenswrapper[4980]: I1206 03:46:16.245754 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/5af0e768-3c9c-4401-ab56-57516cd3170e-host-var-lib-cni-networks-ovn-kubernetes\") pod \"5af0e768-3c9c-4401-ab56-57516cd3170e\" (UID: \"5af0e768-3c9c-4401-ab56-57516cd3170e\") " Dec 06 03:46:16 crc kubenswrapper[4980]: I1206 03:46:16.245776 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/5af0e768-3c9c-4401-ab56-57516cd3170e-etc-openvswitch\") pod \"5af0e768-3c9c-4401-ab56-57516cd3170e\" (UID: \"5af0e768-3c9c-4401-ab56-57516cd3170e\") " Dec 06 03:46:16 crc kubenswrapper[4980]: I1206 03:46:16.245805 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/5af0e768-3c9c-4401-ab56-57516cd3170e-env-overrides\") pod \"5af0e768-3c9c-4401-ab56-57516cd3170e\" (UID: \"5af0e768-3c9c-4401-ab56-57516cd3170e\") " Dec 06 03:46:16 crc kubenswrapper[4980]: I1206 03:46:16.245827 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/5af0e768-3c9c-4401-ab56-57516cd3170e-run-ovn\") pod \"5af0e768-3c9c-4401-ab56-57516cd3170e\" (UID: \"5af0e768-3c9c-4401-ab56-57516cd3170e\") " Dec 06 03:46:16 crc kubenswrapper[4980]: I1206 03:46:16.245851 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/5af0e768-3c9c-4401-ab56-57516cd3170e-log-socket" (OuterVolumeSpecName: "log-socket") pod "5af0e768-3c9c-4401-ab56-57516cd3170e" (UID: "5af0e768-3c9c-4401-ab56-57516cd3170e"). InnerVolumeSpecName "log-socket". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 03:46:16 crc kubenswrapper[4980]: I1206 03:46:16.245872 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/5af0e768-3c9c-4401-ab56-57516cd3170e-run-ovn" (OuterVolumeSpecName: "run-ovn") pod "5af0e768-3c9c-4401-ab56-57516cd3170e" (UID: "5af0e768-3c9c-4401-ab56-57516cd3170e"). InnerVolumeSpecName "run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 03:46:16 crc kubenswrapper[4980]: I1206 03:46:16.245896 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/5af0e768-3c9c-4401-ab56-57516cd3170e-host-var-lib-cni-networks-ovn-kubernetes" (OuterVolumeSpecName: "host-var-lib-cni-networks-ovn-kubernetes") pod "5af0e768-3c9c-4401-ab56-57516cd3170e" (UID: "5af0e768-3c9c-4401-ab56-57516cd3170e"). InnerVolumeSpecName "host-var-lib-cni-networks-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 03:46:16 crc kubenswrapper[4980]: I1206 03:46:16.245916 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/5af0e768-3c9c-4401-ab56-57516cd3170e-etc-openvswitch" (OuterVolumeSpecName: "etc-openvswitch") pod "5af0e768-3c9c-4401-ab56-57516cd3170e" (UID: "5af0e768-3c9c-4401-ab56-57516cd3170e"). InnerVolumeSpecName "etc-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 03:46:16 crc kubenswrapper[4980]: I1206 03:46:16.245988 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/5af0e768-3c9c-4401-ab56-57516cd3170e-host-slash\") pod \"5af0e768-3c9c-4401-ab56-57516cd3170e\" (UID: \"5af0e768-3c9c-4401-ab56-57516cd3170e\") " Dec 06 03:46:16 crc kubenswrapper[4980]: I1206 03:46:16.246032 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/5af0e768-3c9c-4401-ab56-57516cd3170e-host-run-ovn-kubernetes\") pod \"5af0e768-3c9c-4401-ab56-57516cd3170e\" (UID: \"5af0e768-3c9c-4401-ab56-57516cd3170e\") " Dec 06 03:46:16 crc kubenswrapper[4980]: I1206 03:46:16.246065 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/5af0e768-3c9c-4401-ab56-57516cd3170e-systemd-units\") pod \"5af0e768-3c9c-4401-ab56-57516cd3170e\" (UID: \"5af0e768-3c9c-4401-ab56-57516cd3170e\") " Dec 06 03:46:16 crc kubenswrapper[4980]: I1206 03:46:16.246092 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/5af0e768-3c9c-4401-ab56-57516cd3170e-host-slash" (OuterVolumeSpecName: "host-slash") pod "5af0e768-3c9c-4401-ab56-57516cd3170e" (UID: "5af0e768-3c9c-4401-ab56-57516cd3170e"). InnerVolumeSpecName "host-slash". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 03:46:16 crc kubenswrapper[4980]: I1206 03:46:16.246112 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/5af0e768-3c9c-4401-ab56-57516cd3170e-host-run-ovn-kubernetes" (OuterVolumeSpecName: "host-run-ovn-kubernetes") pod "5af0e768-3c9c-4401-ab56-57516cd3170e" (UID: "5af0e768-3c9c-4401-ab56-57516cd3170e"). InnerVolumeSpecName "host-run-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 03:46:16 crc kubenswrapper[4980]: I1206 03:46:16.246129 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/5af0e768-3c9c-4401-ab56-57516cd3170e-systemd-units" (OuterVolumeSpecName: "systemd-units") pod "5af0e768-3c9c-4401-ab56-57516cd3170e" (UID: "5af0e768-3c9c-4401-ab56-57516cd3170e"). InnerVolumeSpecName "systemd-units". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 03:46:16 crc kubenswrapper[4980]: I1206 03:46:16.246367 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5af0e768-3c9c-4401-ab56-57516cd3170e-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "5af0e768-3c9c-4401-ab56-57516cd3170e" (UID: "5af0e768-3c9c-4401-ab56-57516cd3170e"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 03:46:16 crc kubenswrapper[4980]: I1206 03:46:16.246429 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/5af0e768-3c9c-4401-ab56-57516cd3170e-host-kubelet\") pod \"5af0e768-3c9c-4401-ab56-57516cd3170e\" (UID: \"5af0e768-3c9c-4401-ab56-57516cd3170e\") " Dec 06 03:46:16 crc kubenswrapper[4980]: I1206 03:46:16.246464 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/5af0e768-3c9c-4401-ab56-57516cd3170e-ovnkube-config\") pod \"5af0e768-3c9c-4401-ab56-57516cd3170e\" (UID: \"5af0e768-3c9c-4401-ab56-57516cd3170e\") " Dec 06 03:46:16 crc kubenswrapper[4980]: I1206 03:46:16.246482 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/5af0e768-3c9c-4401-ab56-57516cd3170e-node-log\") pod \"5af0e768-3c9c-4401-ab56-57516cd3170e\" (UID: \"5af0e768-3c9c-4401-ab56-57516cd3170e\") " Dec 06 03:46:16 crc kubenswrapper[4980]: I1206 03:46:16.246499 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zfzrb\" (UniqueName: \"kubernetes.io/projected/5af0e768-3c9c-4401-ab56-57516cd3170e-kube-api-access-zfzrb\") pod \"5af0e768-3c9c-4401-ab56-57516cd3170e\" (UID: \"5af0e768-3c9c-4401-ab56-57516cd3170e\") " Dec 06 03:46:16 crc kubenswrapper[4980]: I1206 03:46:16.246540 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/5af0e768-3c9c-4401-ab56-57516cd3170e-run-openvswitch\") pod \"5af0e768-3c9c-4401-ab56-57516cd3170e\" (UID: \"5af0e768-3c9c-4401-ab56-57516cd3170e\") " Dec 06 03:46:16 crc kubenswrapper[4980]: I1206 03:46:16.246562 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/5af0e768-3c9c-4401-ab56-57516cd3170e-host-cni-bin\") pod \"5af0e768-3c9c-4401-ab56-57516cd3170e\" (UID: \"5af0e768-3c9c-4401-ab56-57516cd3170e\") " Dec 06 03:46:16 crc kubenswrapper[4980]: I1206 03:46:16.246598 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/5af0e768-3c9c-4401-ab56-57516cd3170e-ovn-node-metrics-cert\") pod \"5af0e768-3c9c-4401-ab56-57516cd3170e\" (UID: \"5af0e768-3c9c-4401-ab56-57516cd3170e\") " Dec 06 03:46:16 crc kubenswrapper[4980]: I1206 03:46:16.246621 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/5af0e768-3c9c-4401-ab56-57516cd3170e-host-run-netns\") pod \"5af0e768-3c9c-4401-ab56-57516cd3170e\" (UID: \"5af0e768-3c9c-4401-ab56-57516cd3170e\") " Dec 06 03:46:16 crc kubenswrapper[4980]: I1206 03:46:16.246638 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/5af0e768-3c9c-4401-ab56-57516cd3170e-ovnkube-script-lib\") pod \"5af0e768-3c9c-4401-ab56-57516cd3170e\" (UID: \"5af0e768-3c9c-4401-ab56-57516cd3170e\") " Dec 06 03:46:16 crc kubenswrapper[4980]: I1206 03:46:16.246656 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/5af0e768-3c9c-4401-ab56-57516cd3170e-var-lib-openvswitch\") pod \"5af0e768-3c9c-4401-ab56-57516cd3170e\" (UID: \"5af0e768-3c9c-4401-ab56-57516cd3170e\") " Dec 06 03:46:16 crc kubenswrapper[4980]: I1206 03:46:16.246671 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/5af0e768-3c9c-4401-ab56-57516cd3170e-run-systemd\") pod \"5af0e768-3c9c-4401-ab56-57516cd3170e\" (UID: \"5af0e768-3c9c-4401-ab56-57516cd3170e\") " Dec 06 03:46:16 crc kubenswrapper[4980]: I1206 03:46:16.246684 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/5af0e768-3c9c-4401-ab56-57516cd3170e-host-cni-netd\") pod \"5af0e768-3c9c-4401-ab56-57516cd3170e\" (UID: \"5af0e768-3c9c-4401-ab56-57516cd3170e\") " Dec 06 03:46:16 crc kubenswrapper[4980]: I1206 03:46:16.246839 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/e6d35c1e-b567-4b10-9c61-bc4d942f7a24-run-openvswitch\") pod \"ovnkube-node-6hnct\" (UID: \"e6d35c1e-b567-4b10-9c61-bc4d942f7a24\") " pod="openshift-ovn-kubernetes/ovnkube-node-6hnct" Dec 06 03:46:16 crc kubenswrapper[4980]: I1206 03:46:16.246862 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/e6d35c1e-b567-4b10-9c61-bc4d942f7a24-host-slash\") pod \"ovnkube-node-6hnct\" (UID: \"e6d35c1e-b567-4b10-9c61-bc4d942f7a24\") " pod="openshift-ovn-kubernetes/ovnkube-node-6hnct" Dec 06 03:46:16 crc kubenswrapper[4980]: I1206 03:46:16.246883 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/e6d35c1e-b567-4b10-9c61-bc4d942f7a24-host-cni-netd\") pod \"ovnkube-node-6hnct\" (UID: \"e6d35c1e-b567-4b10-9c61-bc4d942f7a24\") " pod="openshift-ovn-kubernetes/ovnkube-node-6hnct" Dec 06 03:46:16 crc kubenswrapper[4980]: I1206 03:46:16.246899 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/e6d35c1e-b567-4b10-9c61-bc4d942f7a24-systemd-units\") pod \"ovnkube-node-6hnct\" (UID: \"e6d35c1e-b567-4b10-9c61-bc4d942f7a24\") " pod="openshift-ovn-kubernetes/ovnkube-node-6hnct" Dec 06 03:46:16 crc kubenswrapper[4980]: I1206 03:46:16.246926 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/e6d35c1e-b567-4b10-9c61-bc4d942f7a24-ovnkube-script-lib\") pod \"ovnkube-node-6hnct\" (UID: \"e6d35c1e-b567-4b10-9c61-bc4d942f7a24\") " pod="openshift-ovn-kubernetes/ovnkube-node-6hnct" Dec 06 03:46:16 crc kubenswrapper[4980]: I1206 03:46:16.246949 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/e6d35c1e-b567-4b10-9c61-bc4d942f7a24-log-socket\") pod \"ovnkube-node-6hnct\" (UID: \"e6d35c1e-b567-4b10-9c61-bc4d942f7a24\") " pod="openshift-ovn-kubernetes/ovnkube-node-6hnct" Dec 06 03:46:16 crc kubenswrapper[4980]: I1206 03:46:16.246985 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/e6d35c1e-b567-4b10-9c61-bc4d942f7a24-host-run-netns\") pod \"ovnkube-node-6hnct\" (UID: \"e6d35c1e-b567-4b10-9c61-bc4d942f7a24\") " pod="openshift-ovn-kubernetes/ovnkube-node-6hnct" Dec 06 03:46:16 crc kubenswrapper[4980]: I1206 03:46:16.247005 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/e6d35c1e-b567-4b10-9c61-bc4d942f7a24-node-log\") pod \"ovnkube-node-6hnct\" (UID: \"e6d35c1e-b567-4b10-9c61-bc4d942f7a24\") " pod="openshift-ovn-kubernetes/ovnkube-node-6hnct" Dec 06 03:46:16 crc kubenswrapper[4980]: I1206 03:46:16.247036 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/e6d35c1e-b567-4b10-9c61-bc4d942f7a24-ovn-node-metrics-cert\") pod \"ovnkube-node-6hnct\" (UID: \"e6d35c1e-b567-4b10-9c61-bc4d942f7a24\") " pod="openshift-ovn-kubernetes/ovnkube-node-6hnct" Dec 06 03:46:16 crc kubenswrapper[4980]: I1206 03:46:16.247052 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/e6d35c1e-b567-4b10-9c61-bc4d942f7a24-run-systemd\") pod \"ovnkube-node-6hnct\" (UID: \"e6d35c1e-b567-4b10-9c61-bc4d942f7a24\") " pod="openshift-ovn-kubernetes/ovnkube-node-6hnct" Dec 06 03:46:16 crc kubenswrapper[4980]: I1206 03:46:16.247081 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/e6d35c1e-b567-4b10-9c61-bc4d942f7a24-etc-openvswitch\") pod \"ovnkube-node-6hnct\" (UID: \"e6d35c1e-b567-4b10-9c61-bc4d942f7a24\") " pod="openshift-ovn-kubernetes/ovnkube-node-6hnct" Dec 06 03:46:16 crc kubenswrapper[4980]: I1206 03:46:16.247098 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/e6d35c1e-b567-4b10-9c61-bc4d942f7a24-host-run-ovn-kubernetes\") pod \"ovnkube-node-6hnct\" (UID: \"e6d35c1e-b567-4b10-9c61-bc4d942f7a24\") " pod="openshift-ovn-kubernetes/ovnkube-node-6hnct" Dec 06 03:46:16 crc kubenswrapper[4980]: I1206 03:46:16.247114 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-69xkd\" (UniqueName: \"kubernetes.io/projected/e6d35c1e-b567-4b10-9c61-bc4d942f7a24-kube-api-access-69xkd\") pod \"ovnkube-node-6hnct\" (UID: \"e6d35c1e-b567-4b10-9c61-bc4d942f7a24\") " pod="openshift-ovn-kubernetes/ovnkube-node-6hnct" Dec 06 03:46:16 crc kubenswrapper[4980]: I1206 03:46:16.247132 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/e6d35c1e-b567-4b10-9c61-bc4d942f7a24-host-kubelet\") pod \"ovnkube-node-6hnct\" (UID: \"e6d35c1e-b567-4b10-9c61-bc4d942f7a24\") " pod="openshift-ovn-kubernetes/ovnkube-node-6hnct" Dec 06 03:46:16 crc kubenswrapper[4980]: I1206 03:46:16.247166 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/e6d35c1e-b567-4b10-9c61-bc4d942f7a24-run-ovn\") pod \"ovnkube-node-6hnct\" (UID: \"e6d35c1e-b567-4b10-9c61-bc4d942f7a24\") " pod="openshift-ovn-kubernetes/ovnkube-node-6hnct" Dec 06 03:46:16 crc kubenswrapper[4980]: I1206 03:46:16.247186 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/e6d35c1e-b567-4b10-9c61-bc4d942f7a24-host-cni-bin\") pod \"ovnkube-node-6hnct\" (UID: \"e6d35c1e-b567-4b10-9c61-bc4d942f7a24\") " pod="openshift-ovn-kubernetes/ovnkube-node-6hnct" Dec 06 03:46:16 crc kubenswrapper[4980]: I1206 03:46:16.247201 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/e6d35c1e-b567-4b10-9c61-bc4d942f7a24-ovnkube-config\") pod \"ovnkube-node-6hnct\" (UID: \"e6d35c1e-b567-4b10-9c61-bc4d942f7a24\") " pod="openshift-ovn-kubernetes/ovnkube-node-6hnct" Dec 06 03:46:16 crc kubenswrapper[4980]: I1206 03:46:16.247216 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/e6d35c1e-b567-4b10-9c61-bc4d942f7a24-var-lib-openvswitch\") pod \"ovnkube-node-6hnct\" (UID: \"e6d35c1e-b567-4b10-9c61-bc4d942f7a24\") " pod="openshift-ovn-kubernetes/ovnkube-node-6hnct" Dec 06 03:46:16 crc kubenswrapper[4980]: I1206 03:46:16.247231 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/e6d35c1e-b567-4b10-9c61-bc4d942f7a24-env-overrides\") pod \"ovnkube-node-6hnct\" (UID: \"e6d35c1e-b567-4b10-9c61-bc4d942f7a24\") " pod="openshift-ovn-kubernetes/ovnkube-node-6hnct" Dec 06 03:46:16 crc kubenswrapper[4980]: I1206 03:46:16.247251 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/e6d35c1e-b567-4b10-9c61-bc4d942f7a24-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-6hnct\" (UID: \"e6d35c1e-b567-4b10-9c61-bc4d942f7a24\") " pod="openshift-ovn-kubernetes/ovnkube-node-6hnct" Dec 06 03:46:16 crc kubenswrapper[4980]: I1206 03:46:16.247282 4980 reconciler_common.go:293] "Volume detached for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/5af0e768-3c9c-4401-ab56-57516cd3170e-log-socket\") on node \"crc\" DevicePath \"\"" Dec 06 03:46:16 crc kubenswrapper[4980]: I1206 03:46:16.247293 4980 reconciler_common.go:293] "Volume detached for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/5af0e768-3c9c-4401-ab56-57516cd3170e-host-var-lib-cni-networks-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Dec 06 03:46:16 crc kubenswrapper[4980]: I1206 03:46:16.247303 4980 reconciler_common.go:293] "Volume detached for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/5af0e768-3c9c-4401-ab56-57516cd3170e-etc-openvswitch\") on node \"crc\" DevicePath \"\"" Dec 06 03:46:16 crc kubenswrapper[4980]: I1206 03:46:16.247313 4980 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/5af0e768-3c9c-4401-ab56-57516cd3170e-env-overrides\") on node \"crc\" DevicePath \"\"" Dec 06 03:46:16 crc kubenswrapper[4980]: I1206 03:46:16.247323 4980 reconciler_common.go:293] "Volume detached for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/5af0e768-3c9c-4401-ab56-57516cd3170e-run-ovn\") on node \"crc\" DevicePath \"\"" Dec 06 03:46:16 crc kubenswrapper[4980]: I1206 03:46:16.247332 4980 reconciler_common.go:293] "Volume detached for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/5af0e768-3c9c-4401-ab56-57516cd3170e-host-slash\") on node \"crc\" DevicePath \"\"" Dec 06 03:46:16 crc kubenswrapper[4980]: I1206 03:46:16.247341 4980 reconciler_common.go:293] "Volume detached for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/5af0e768-3c9c-4401-ab56-57516cd3170e-host-run-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Dec 06 03:46:16 crc kubenswrapper[4980]: I1206 03:46:16.247349 4980 reconciler_common.go:293] "Volume detached for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/5af0e768-3c9c-4401-ab56-57516cd3170e-systemd-units\") on node \"crc\" DevicePath \"\"" Dec 06 03:46:16 crc kubenswrapper[4980]: I1206 03:46:16.247381 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/5af0e768-3c9c-4401-ab56-57516cd3170e-run-openvswitch" (OuterVolumeSpecName: "run-openvswitch") pod "5af0e768-3c9c-4401-ab56-57516cd3170e" (UID: "5af0e768-3c9c-4401-ab56-57516cd3170e"). InnerVolumeSpecName "run-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 03:46:16 crc kubenswrapper[4980]: I1206 03:46:16.247399 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/5af0e768-3c9c-4401-ab56-57516cd3170e-host-cni-bin" (OuterVolumeSpecName: "host-cni-bin") pod "5af0e768-3c9c-4401-ab56-57516cd3170e" (UID: "5af0e768-3c9c-4401-ab56-57516cd3170e"). InnerVolumeSpecName "host-cni-bin". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 03:46:16 crc kubenswrapper[4980]: I1206 03:46:16.247938 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/5af0e768-3c9c-4401-ab56-57516cd3170e-host-kubelet" (OuterVolumeSpecName: "host-kubelet") pod "5af0e768-3c9c-4401-ab56-57516cd3170e" (UID: "5af0e768-3c9c-4401-ab56-57516cd3170e"). InnerVolumeSpecName "host-kubelet". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 03:46:16 crc kubenswrapper[4980]: I1206 03:46:16.248012 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/5af0e768-3c9c-4401-ab56-57516cd3170e-node-log" (OuterVolumeSpecName: "node-log") pod "5af0e768-3c9c-4401-ab56-57516cd3170e" (UID: "5af0e768-3c9c-4401-ab56-57516cd3170e"). InnerVolumeSpecName "node-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 03:46:16 crc kubenswrapper[4980]: I1206 03:46:16.248070 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/5af0e768-3c9c-4401-ab56-57516cd3170e-var-lib-openvswitch" (OuterVolumeSpecName: "var-lib-openvswitch") pod "5af0e768-3c9c-4401-ab56-57516cd3170e" (UID: "5af0e768-3c9c-4401-ab56-57516cd3170e"). InnerVolumeSpecName "var-lib-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 03:46:16 crc kubenswrapper[4980]: I1206 03:46:16.248094 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/5af0e768-3c9c-4401-ab56-57516cd3170e-host-run-netns" (OuterVolumeSpecName: "host-run-netns") pod "5af0e768-3c9c-4401-ab56-57516cd3170e" (UID: "5af0e768-3c9c-4401-ab56-57516cd3170e"). InnerVolumeSpecName "host-run-netns". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 03:46:16 crc kubenswrapper[4980]: I1206 03:46:16.248408 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5af0e768-3c9c-4401-ab56-57516cd3170e-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "5af0e768-3c9c-4401-ab56-57516cd3170e" (UID: "5af0e768-3c9c-4401-ab56-57516cd3170e"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 03:46:16 crc kubenswrapper[4980]: I1206 03:46:16.248441 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/5af0e768-3c9c-4401-ab56-57516cd3170e-host-cni-netd" (OuterVolumeSpecName: "host-cni-netd") pod "5af0e768-3c9c-4401-ab56-57516cd3170e" (UID: "5af0e768-3c9c-4401-ab56-57516cd3170e"). InnerVolumeSpecName "host-cni-netd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 03:46:16 crc kubenswrapper[4980]: I1206 03:46:16.248603 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5af0e768-3c9c-4401-ab56-57516cd3170e-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "5af0e768-3c9c-4401-ab56-57516cd3170e" (UID: "5af0e768-3c9c-4401-ab56-57516cd3170e"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 03:46:16 crc kubenswrapper[4980]: I1206 03:46:16.252413 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5af0e768-3c9c-4401-ab56-57516cd3170e-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "5af0e768-3c9c-4401-ab56-57516cd3170e" (UID: "5af0e768-3c9c-4401-ab56-57516cd3170e"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 03:46:16 crc kubenswrapper[4980]: I1206 03:46:16.253040 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5af0e768-3c9c-4401-ab56-57516cd3170e-kube-api-access-zfzrb" (OuterVolumeSpecName: "kube-api-access-zfzrb") pod "5af0e768-3c9c-4401-ab56-57516cd3170e" (UID: "5af0e768-3c9c-4401-ab56-57516cd3170e"). InnerVolumeSpecName "kube-api-access-zfzrb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 03:46:16 crc kubenswrapper[4980]: I1206 03:46:16.263848 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/5af0e768-3c9c-4401-ab56-57516cd3170e-run-systemd" (OuterVolumeSpecName: "run-systemd") pod "5af0e768-3c9c-4401-ab56-57516cd3170e" (UID: "5af0e768-3c9c-4401-ab56-57516cd3170e"). InnerVolumeSpecName "run-systemd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 03:46:16 crc kubenswrapper[4980]: I1206 03:46:16.347735 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/e6d35c1e-b567-4b10-9c61-bc4d942f7a24-log-socket\") pod \"ovnkube-node-6hnct\" (UID: \"e6d35c1e-b567-4b10-9c61-bc4d942f7a24\") " pod="openshift-ovn-kubernetes/ovnkube-node-6hnct" Dec 06 03:46:16 crc kubenswrapper[4980]: I1206 03:46:16.347804 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/e6d35c1e-b567-4b10-9c61-bc4d942f7a24-host-run-netns\") pod \"ovnkube-node-6hnct\" (UID: \"e6d35c1e-b567-4b10-9c61-bc4d942f7a24\") " pod="openshift-ovn-kubernetes/ovnkube-node-6hnct" Dec 06 03:46:16 crc kubenswrapper[4980]: I1206 03:46:16.347824 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/e6d35c1e-b567-4b10-9c61-bc4d942f7a24-node-log\") pod \"ovnkube-node-6hnct\" (UID: \"e6d35c1e-b567-4b10-9c61-bc4d942f7a24\") " pod="openshift-ovn-kubernetes/ovnkube-node-6hnct" Dec 06 03:46:16 crc kubenswrapper[4980]: I1206 03:46:16.347890 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/e6d35c1e-b567-4b10-9c61-bc4d942f7a24-host-run-netns\") pod \"ovnkube-node-6hnct\" (UID: \"e6d35c1e-b567-4b10-9c61-bc4d942f7a24\") " pod="openshift-ovn-kubernetes/ovnkube-node-6hnct" Dec 06 03:46:16 crc kubenswrapper[4980]: I1206 03:46:16.347923 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/e6d35c1e-b567-4b10-9c61-bc4d942f7a24-ovn-node-metrics-cert\") pod \"ovnkube-node-6hnct\" (UID: \"e6d35c1e-b567-4b10-9c61-bc4d942f7a24\") " pod="openshift-ovn-kubernetes/ovnkube-node-6hnct" Dec 06 03:46:16 crc kubenswrapper[4980]: I1206 03:46:16.347939 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/e6d35c1e-b567-4b10-9c61-bc4d942f7a24-run-systemd\") pod \"ovnkube-node-6hnct\" (UID: \"e6d35c1e-b567-4b10-9c61-bc4d942f7a24\") " pod="openshift-ovn-kubernetes/ovnkube-node-6hnct" Dec 06 03:46:16 crc kubenswrapper[4980]: I1206 03:46:16.347956 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/e6d35c1e-b567-4b10-9c61-bc4d942f7a24-etc-openvswitch\") pod \"ovnkube-node-6hnct\" (UID: \"e6d35c1e-b567-4b10-9c61-bc4d942f7a24\") " pod="openshift-ovn-kubernetes/ovnkube-node-6hnct" Dec 06 03:46:16 crc kubenswrapper[4980]: I1206 03:46:16.347918 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/e6d35c1e-b567-4b10-9c61-bc4d942f7a24-node-log\") pod \"ovnkube-node-6hnct\" (UID: \"e6d35c1e-b567-4b10-9c61-bc4d942f7a24\") " pod="openshift-ovn-kubernetes/ovnkube-node-6hnct" Dec 06 03:46:16 crc kubenswrapper[4980]: I1206 03:46:16.348021 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-69xkd\" (UniqueName: \"kubernetes.io/projected/e6d35c1e-b567-4b10-9c61-bc4d942f7a24-kube-api-access-69xkd\") pod \"ovnkube-node-6hnct\" (UID: \"e6d35c1e-b567-4b10-9c61-bc4d942f7a24\") " pod="openshift-ovn-kubernetes/ovnkube-node-6hnct" Dec 06 03:46:16 crc kubenswrapper[4980]: I1206 03:46:16.348039 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/e6d35c1e-b567-4b10-9c61-bc4d942f7a24-host-run-ovn-kubernetes\") pod \"ovnkube-node-6hnct\" (UID: \"e6d35c1e-b567-4b10-9c61-bc4d942f7a24\") " pod="openshift-ovn-kubernetes/ovnkube-node-6hnct" Dec 06 03:46:16 crc kubenswrapper[4980]: I1206 03:46:16.348071 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/e6d35c1e-b567-4b10-9c61-bc4d942f7a24-run-systemd\") pod \"ovnkube-node-6hnct\" (UID: \"e6d35c1e-b567-4b10-9c61-bc4d942f7a24\") " pod="openshift-ovn-kubernetes/ovnkube-node-6hnct" Dec 06 03:46:16 crc kubenswrapper[4980]: I1206 03:46:16.348207 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/e6d35c1e-b567-4b10-9c61-bc4d942f7a24-etc-openvswitch\") pod \"ovnkube-node-6hnct\" (UID: \"e6d35c1e-b567-4b10-9c61-bc4d942f7a24\") " pod="openshift-ovn-kubernetes/ovnkube-node-6hnct" Dec 06 03:46:16 crc kubenswrapper[4980]: I1206 03:46:16.348354 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/e6d35c1e-b567-4b10-9c61-bc4d942f7a24-host-run-ovn-kubernetes\") pod \"ovnkube-node-6hnct\" (UID: \"e6d35c1e-b567-4b10-9c61-bc4d942f7a24\") " pod="openshift-ovn-kubernetes/ovnkube-node-6hnct" Dec 06 03:46:16 crc kubenswrapper[4980]: I1206 03:46:16.348363 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/e6d35c1e-b567-4b10-9c61-bc4d942f7a24-host-kubelet\") pod \"ovnkube-node-6hnct\" (UID: \"e6d35c1e-b567-4b10-9c61-bc4d942f7a24\") " pod="openshift-ovn-kubernetes/ovnkube-node-6hnct" Dec 06 03:46:16 crc kubenswrapper[4980]: I1206 03:46:16.348446 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/e6d35c1e-b567-4b10-9c61-bc4d942f7a24-host-kubelet\") pod \"ovnkube-node-6hnct\" (UID: \"e6d35c1e-b567-4b10-9c61-bc4d942f7a24\") " pod="openshift-ovn-kubernetes/ovnkube-node-6hnct" Dec 06 03:46:16 crc kubenswrapper[4980]: I1206 03:46:16.348487 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/e6d35c1e-b567-4b10-9c61-bc4d942f7a24-host-cni-bin\") pod \"ovnkube-node-6hnct\" (UID: \"e6d35c1e-b567-4b10-9c61-bc4d942f7a24\") " pod="openshift-ovn-kubernetes/ovnkube-node-6hnct" Dec 06 03:46:16 crc kubenswrapper[4980]: I1206 03:46:16.348555 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/e6d35c1e-b567-4b10-9c61-bc4d942f7a24-ovnkube-config\") pod \"ovnkube-node-6hnct\" (UID: \"e6d35c1e-b567-4b10-9c61-bc4d942f7a24\") " pod="openshift-ovn-kubernetes/ovnkube-node-6hnct" Dec 06 03:46:16 crc kubenswrapper[4980]: I1206 03:46:16.348580 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/e6d35c1e-b567-4b10-9c61-bc4d942f7a24-run-ovn\") pod \"ovnkube-node-6hnct\" (UID: \"e6d35c1e-b567-4b10-9c61-bc4d942f7a24\") " pod="openshift-ovn-kubernetes/ovnkube-node-6hnct" Dec 06 03:46:16 crc kubenswrapper[4980]: I1206 03:46:16.348624 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/e6d35c1e-b567-4b10-9c61-bc4d942f7a24-env-overrides\") pod \"ovnkube-node-6hnct\" (UID: \"e6d35c1e-b567-4b10-9c61-bc4d942f7a24\") " pod="openshift-ovn-kubernetes/ovnkube-node-6hnct" Dec 06 03:46:16 crc kubenswrapper[4980]: I1206 03:46:16.348650 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/e6d35c1e-b567-4b10-9c61-bc4d942f7a24-var-lib-openvswitch\") pod \"ovnkube-node-6hnct\" (UID: \"e6d35c1e-b567-4b10-9c61-bc4d942f7a24\") " pod="openshift-ovn-kubernetes/ovnkube-node-6hnct" Dec 06 03:46:16 crc kubenswrapper[4980]: I1206 03:46:16.348675 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/e6d35c1e-b567-4b10-9c61-bc4d942f7a24-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-6hnct\" (UID: \"e6d35c1e-b567-4b10-9c61-bc4d942f7a24\") " pod="openshift-ovn-kubernetes/ovnkube-node-6hnct" Dec 06 03:46:16 crc kubenswrapper[4980]: I1206 03:46:16.348738 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/e6d35c1e-b567-4b10-9c61-bc4d942f7a24-log-socket\") pod \"ovnkube-node-6hnct\" (UID: \"e6d35c1e-b567-4b10-9c61-bc4d942f7a24\") " pod="openshift-ovn-kubernetes/ovnkube-node-6hnct" Dec 06 03:46:16 crc kubenswrapper[4980]: I1206 03:46:16.348811 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/e6d35c1e-b567-4b10-9c61-bc4d942f7a24-run-ovn\") pod \"ovnkube-node-6hnct\" (UID: \"e6d35c1e-b567-4b10-9c61-bc4d942f7a24\") " pod="openshift-ovn-kubernetes/ovnkube-node-6hnct" Dec 06 03:46:16 crc kubenswrapper[4980]: I1206 03:46:16.348820 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/e6d35c1e-b567-4b10-9c61-bc4d942f7a24-run-openvswitch\") pod \"ovnkube-node-6hnct\" (UID: \"e6d35c1e-b567-4b10-9c61-bc4d942f7a24\") " pod="openshift-ovn-kubernetes/ovnkube-node-6hnct" Dec 06 03:46:16 crc kubenswrapper[4980]: I1206 03:46:16.348571 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/e6d35c1e-b567-4b10-9c61-bc4d942f7a24-host-cni-bin\") pod \"ovnkube-node-6hnct\" (UID: \"e6d35c1e-b567-4b10-9c61-bc4d942f7a24\") " pod="openshift-ovn-kubernetes/ovnkube-node-6hnct" Dec 06 03:46:16 crc kubenswrapper[4980]: I1206 03:46:16.348751 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/e6d35c1e-b567-4b10-9c61-bc4d942f7a24-run-openvswitch\") pod \"ovnkube-node-6hnct\" (UID: \"e6d35c1e-b567-4b10-9c61-bc4d942f7a24\") " pod="openshift-ovn-kubernetes/ovnkube-node-6hnct" Dec 06 03:46:16 crc kubenswrapper[4980]: I1206 03:46:16.348883 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/e6d35c1e-b567-4b10-9c61-bc4d942f7a24-host-slash\") pod \"ovnkube-node-6hnct\" (UID: \"e6d35c1e-b567-4b10-9c61-bc4d942f7a24\") " pod="openshift-ovn-kubernetes/ovnkube-node-6hnct" Dec 06 03:46:16 crc kubenswrapper[4980]: I1206 03:46:16.348918 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/e6d35c1e-b567-4b10-9c61-bc4d942f7a24-host-cni-netd\") pod \"ovnkube-node-6hnct\" (UID: \"e6d35c1e-b567-4b10-9c61-bc4d942f7a24\") " pod="openshift-ovn-kubernetes/ovnkube-node-6hnct" Dec 06 03:46:16 crc kubenswrapper[4980]: I1206 03:46:16.348941 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/e6d35c1e-b567-4b10-9c61-bc4d942f7a24-systemd-units\") pod \"ovnkube-node-6hnct\" (UID: \"e6d35c1e-b567-4b10-9c61-bc4d942f7a24\") " pod="openshift-ovn-kubernetes/ovnkube-node-6hnct" Dec 06 03:46:16 crc kubenswrapper[4980]: I1206 03:46:16.348966 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/e6d35c1e-b567-4b10-9c61-bc4d942f7a24-ovnkube-script-lib\") pod \"ovnkube-node-6hnct\" (UID: \"e6d35c1e-b567-4b10-9c61-bc4d942f7a24\") " pod="openshift-ovn-kubernetes/ovnkube-node-6hnct" Dec 06 03:46:16 crc kubenswrapper[4980]: I1206 03:46:16.349060 4980 reconciler_common.go:293] "Volume detached for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/5af0e768-3c9c-4401-ab56-57516cd3170e-host-cni-bin\") on node \"crc\" DevicePath \"\"" Dec 06 03:46:16 crc kubenswrapper[4980]: I1206 03:46:16.349073 4980 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/5af0e768-3c9c-4401-ab56-57516cd3170e-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Dec 06 03:46:16 crc kubenswrapper[4980]: I1206 03:46:16.349103 4980 reconciler_common.go:293] "Volume detached for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/5af0e768-3c9c-4401-ab56-57516cd3170e-host-run-netns\") on node \"crc\" DevicePath \"\"" Dec 06 03:46:16 crc kubenswrapper[4980]: I1206 03:46:16.349117 4980 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/5af0e768-3c9c-4401-ab56-57516cd3170e-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Dec 06 03:46:16 crc kubenswrapper[4980]: I1206 03:46:16.349129 4980 reconciler_common.go:293] "Volume detached for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/5af0e768-3c9c-4401-ab56-57516cd3170e-var-lib-openvswitch\") on node \"crc\" DevicePath \"\"" Dec 06 03:46:16 crc kubenswrapper[4980]: I1206 03:46:16.349142 4980 reconciler_common.go:293] "Volume detached for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/5af0e768-3c9c-4401-ab56-57516cd3170e-run-systemd\") on node \"crc\" DevicePath \"\"" Dec 06 03:46:16 crc kubenswrapper[4980]: I1206 03:46:16.349153 4980 reconciler_common.go:293] "Volume detached for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/5af0e768-3c9c-4401-ab56-57516cd3170e-host-cni-netd\") on node \"crc\" DevicePath \"\"" Dec 06 03:46:16 crc kubenswrapper[4980]: I1206 03:46:16.349167 4980 reconciler_common.go:293] "Volume detached for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/5af0e768-3c9c-4401-ab56-57516cd3170e-host-kubelet\") on node \"crc\" DevicePath \"\"" Dec 06 03:46:16 crc kubenswrapper[4980]: I1206 03:46:16.349182 4980 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/5af0e768-3c9c-4401-ab56-57516cd3170e-ovnkube-config\") on node \"crc\" DevicePath \"\"" Dec 06 03:46:16 crc kubenswrapper[4980]: I1206 03:46:16.349193 4980 reconciler_common.go:293] "Volume detached for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/5af0e768-3c9c-4401-ab56-57516cd3170e-node-log\") on node \"crc\" DevicePath \"\"" Dec 06 03:46:16 crc kubenswrapper[4980]: I1206 03:46:16.349204 4980 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zfzrb\" (UniqueName: \"kubernetes.io/projected/5af0e768-3c9c-4401-ab56-57516cd3170e-kube-api-access-zfzrb\") on node \"crc\" DevicePath \"\"" Dec 06 03:46:16 crc kubenswrapper[4980]: I1206 03:46:16.349215 4980 reconciler_common.go:293] "Volume detached for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/5af0e768-3c9c-4401-ab56-57516cd3170e-run-openvswitch\") on node \"crc\" DevicePath \"\"" Dec 06 03:46:16 crc kubenswrapper[4980]: I1206 03:46:16.349561 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/e6d35c1e-b567-4b10-9c61-bc4d942f7a24-env-overrides\") pod \"ovnkube-node-6hnct\" (UID: \"e6d35c1e-b567-4b10-9c61-bc4d942f7a24\") " pod="openshift-ovn-kubernetes/ovnkube-node-6hnct" Dec 06 03:46:16 crc kubenswrapper[4980]: I1206 03:46:16.349634 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/e6d35c1e-b567-4b10-9c61-bc4d942f7a24-var-lib-openvswitch\") pod \"ovnkube-node-6hnct\" (UID: \"e6d35c1e-b567-4b10-9c61-bc4d942f7a24\") " pod="openshift-ovn-kubernetes/ovnkube-node-6hnct" Dec 06 03:46:16 crc kubenswrapper[4980]: I1206 03:46:16.349666 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/e6d35c1e-b567-4b10-9c61-bc4d942f7a24-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-6hnct\" (UID: \"e6d35c1e-b567-4b10-9c61-bc4d942f7a24\") " pod="openshift-ovn-kubernetes/ovnkube-node-6hnct" Dec 06 03:46:16 crc kubenswrapper[4980]: I1206 03:46:16.349722 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/e6d35c1e-b567-4b10-9c61-bc4d942f7a24-host-cni-netd\") pod \"ovnkube-node-6hnct\" (UID: \"e6d35c1e-b567-4b10-9c61-bc4d942f7a24\") " pod="openshift-ovn-kubernetes/ovnkube-node-6hnct" Dec 06 03:46:16 crc kubenswrapper[4980]: I1206 03:46:16.349752 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/e6d35c1e-b567-4b10-9c61-bc4d942f7a24-host-slash\") pod \"ovnkube-node-6hnct\" (UID: \"e6d35c1e-b567-4b10-9c61-bc4d942f7a24\") " pod="openshift-ovn-kubernetes/ovnkube-node-6hnct" Dec 06 03:46:16 crc kubenswrapper[4980]: I1206 03:46:16.349798 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/e6d35c1e-b567-4b10-9c61-bc4d942f7a24-systemd-units\") pod \"ovnkube-node-6hnct\" (UID: \"e6d35c1e-b567-4b10-9c61-bc4d942f7a24\") " pod="openshift-ovn-kubernetes/ovnkube-node-6hnct" Dec 06 03:46:16 crc kubenswrapper[4980]: I1206 03:46:16.349833 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/e6d35c1e-b567-4b10-9c61-bc4d942f7a24-ovnkube-config\") pod \"ovnkube-node-6hnct\" (UID: \"e6d35c1e-b567-4b10-9c61-bc4d942f7a24\") " pod="openshift-ovn-kubernetes/ovnkube-node-6hnct" Dec 06 03:46:16 crc kubenswrapper[4980]: I1206 03:46:16.349914 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/e6d35c1e-b567-4b10-9c61-bc4d942f7a24-ovnkube-script-lib\") pod \"ovnkube-node-6hnct\" (UID: \"e6d35c1e-b567-4b10-9c61-bc4d942f7a24\") " pod="openshift-ovn-kubernetes/ovnkube-node-6hnct" Dec 06 03:46:16 crc kubenswrapper[4980]: I1206 03:46:16.351318 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/e6d35c1e-b567-4b10-9c61-bc4d942f7a24-ovn-node-metrics-cert\") pod \"ovnkube-node-6hnct\" (UID: \"e6d35c1e-b567-4b10-9c61-bc4d942f7a24\") " pod="openshift-ovn-kubernetes/ovnkube-node-6hnct" Dec 06 03:46:16 crc kubenswrapper[4980]: I1206 03:46:16.363820 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-69xkd\" (UniqueName: \"kubernetes.io/projected/e6d35c1e-b567-4b10-9c61-bc4d942f7a24-kube-api-access-69xkd\") pod \"ovnkube-node-6hnct\" (UID: \"e6d35c1e-b567-4b10-9c61-bc4d942f7a24\") " pod="openshift-ovn-kubernetes/ovnkube-node-6hnct" Dec 06 03:46:16 crc kubenswrapper[4980]: I1206 03:46:16.464939 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-6hnct" Dec 06 03:46:16 crc kubenswrapper[4980]: I1206 03:46:16.632618 4980 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-kq4z8_f96c1f6c-55e8-4d86-bff3-5381581b7022/kube-multus/2.log" Dec 06 03:46:16 crc kubenswrapper[4980]: I1206 03:46:16.633442 4980 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-kq4z8_f96c1f6c-55e8-4d86-bff3-5381581b7022/kube-multus/1.log" Dec 06 03:46:16 crc kubenswrapper[4980]: I1206 03:46:16.633495 4980 generic.go:334] "Generic (PLEG): container finished" podID="f96c1f6c-55e8-4d86-bff3-5381581b7022" containerID="f894aadbd8dc641edcede1ab30092ef8087c9314de751da017e522e14405996a" exitCode=2 Dec 06 03:46:16 crc kubenswrapper[4980]: I1206 03:46:16.633611 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-kq4z8" event={"ID":"f96c1f6c-55e8-4d86-bff3-5381581b7022","Type":"ContainerDied","Data":"f894aadbd8dc641edcede1ab30092ef8087c9314de751da017e522e14405996a"} Dec 06 03:46:16 crc kubenswrapper[4980]: I1206 03:46:16.633672 4980 scope.go:117] "RemoveContainer" containerID="ea7f0bc23247b5effe7f0250582c8694bbf0e9a2d185a844773665a0a48f350e" Dec 06 03:46:16 crc kubenswrapper[4980]: I1206 03:46:16.634348 4980 scope.go:117] "RemoveContainer" containerID="f894aadbd8dc641edcede1ab30092ef8087c9314de751da017e522e14405996a" Dec 06 03:46:16 crc kubenswrapper[4980]: I1206 03:46:16.660424 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6hnct" event={"ID":"e6d35c1e-b567-4b10-9c61-bc4d942f7a24","Type":"ContainerStarted","Data":"65e9b479552f0e278bcdbc379a8e7cdce118ec9b5169fd870e92a23d5f57353c"} Dec 06 03:46:16 crc kubenswrapper[4980]: I1206 03:46:16.664805 4980 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-5w4fr_5af0e768-3c9c-4401-ab56-57516cd3170e/ovnkube-controller/3.log" Dec 06 03:46:16 crc kubenswrapper[4980]: I1206 03:46:16.689554 4980 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-5w4fr_5af0e768-3c9c-4401-ab56-57516cd3170e/ovn-acl-logging/0.log" Dec 06 03:46:16 crc kubenswrapper[4980]: I1206 03:46:16.691361 4980 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-5w4fr_5af0e768-3c9c-4401-ab56-57516cd3170e/ovn-controller/0.log" Dec 06 03:46:16 crc kubenswrapper[4980]: I1206 03:46:16.692256 4980 generic.go:334] "Generic (PLEG): container finished" podID="5af0e768-3c9c-4401-ab56-57516cd3170e" containerID="2d3ca7fa81b93eca00b5d90c8b2685ff607a8156fe318cf99c5296fd65a5be75" exitCode=0 Dec 06 03:46:16 crc kubenswrapper[4980]: I1206 03:46:16.692297 4980 generic.go:334] "Generic (PLEG): container finished" podID="5af0e768-3c9c-4401-ab56-57516cd3170e" containerID="fe261f4322b85702181d7473ab2a9a4d2d4b869d586c3091f811b12a81a63aae" exitCode=0 Dec 06 03:46:16 crc kubenswrapper[4980]: I1206 03:46:16.692305 4980 generic.go:334] "Generic (PLEG): container finished" podID="5af0e768-3c9c-4401-ab56-57516cd3170e" containerID="2d620685574db7594132ae24a91ec0a93dce20a9a289540deebb39650b8a1816" exitCode=0 Dec 06 03:46:16 crc kubenswrapper[4980]: I1206 03:46:16.692313 4980 generic.go:334] "Generic (PLEG): container finished" podID="5af0e768-3c9c-4401-ab56-57516cd3170e" containerID="c67ec8f9c1764fcdf215d52822c2d4f0437dc7f056f05d708767121f6c1f3515" exitCode=0 Dec 06 03:46:16 crc kubenswrapper[4980]: I1206 03:46:16.692319 4980 generic.go:334] "Generic (PLEG): container finished" podID="5af0e768-3c9c-4401-ab56-57516cd3170e" containerID="6342903b318317ee8d84c2f5d467cb25c811a789be0d6ab0e0cba37b5611e16d" exitCode=0 Dec 06 03:46:16 crc kubenswrapper[4980]: I1206 03:46:16.692325 4980 generic.go:334] "Generic (PLEG): container finished" podID="5af0e768-3c9c-4401-ab56-57516cd3170e" containerID="768b8e1fe1a00a8467d744da5ad3f61006b0a3612c96f8e4d0ee6fcb93195f9f" exitCode=0 Dec 06 03:46:16 crc kubenswrapper[4980]: I1206 03:46:16.692331 4980 generic.go:334] "Generic (PLEG): container finished" podID="5af0e768-3c9c-4401-ab56-57516cd3170e" containerID="f35beae5339eeaf26494ed678e3c1bd02349eae2a4b4615325b17923f3b653b9" exitCode=143 Dec 06 03:46:16 crc kubenswrapper[4980]: I1206 03:46:16.692339 4980 generic.go:334] "Generic (PLEG): container finished" podID="5af0e768-3c9c-4401-ab56-57516cd3170e" containerID="a940ec6ee523313728f7e8ae934b7e0b0cff82771e3e64723309db1fae013c3a" exitCode=143 Dec 06 03:46:16 crc kubenswrapper[4980]: I1206 03:46:16.692361 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5w4fr" event={"ID":"5af0e768-3c9c-4401-ab56-57516cd3170e","Type":"ContainerDied","Data":"2d3ca7fa81b93eca00b5d90c8b2685ff607a8156fe318cf99c5296fd65a5be75"} Dec 06 03:46:16 crc kubenswrapper[4980]: I1206 03:46:16.692388 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5w4fr" event={"ID":"5af0e768-3c9c-4401-ab56-57516cd3170e","Type":"ContainerDied","Data":"fe261f4322b85702181d7473ab2a9a4d2d4b869d586c3091f811b12a81a63aae"} Dec 06 03:46:16 crc kubenswrapper[4980]: I1206 03:46:16.692388 4980 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-5w4fr" Dec 06 03:46:16 crc kubenswrapper[4980]: I1206 03:46:16.692399 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5w4fr" event={"ID":"5af0e768-3c9c-4401-ab56-57516cd3170e","Type":"ContainerDied","Data":"2d620685574db7594132ae24a91ec0a93dce20a9a289540deebb39650b8a1816"} Dec 06 03:46:16 crc kubenswrapper[4980]: I1206 03:46:16.693575 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5w4fr" event={"ID":"5af0e768-3c9c-4401-ab56-57516cd3170e","Type":"ContainerDied","Data":"c67ec8f9c1764fcdf215d52822c2d4f0437dc7f056f05d708767121f6c1f3515"} Dec 06 03:46:16 crc kubenswrapper[4980]: I1206 03:46:16.693798 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5w4fr" event={"ID":"5af0e768-3c9c-4401-ab56-57516cd3170e","Type":"ContainerDied","Data":"6342903b318317ee8d84c2f5d467cb25c811a789be0d6ab0e0cba37b5611e16d"} Dec 06 03:46:16 crc kubenswrapper[4980]: I1206 03:46:16.694937 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5w4fr" event={"ID":"5af0e768-3c9c-4401-ab56-57516cd3170e","Type":"ContainerDied","Data":"768b8e1fe1a00a8467d744da5ad3f61006b0a3612c96f8e4d0ee6fcb93195f9f"} Dec 06 03:46:16 crc kubenswrapper[4980]: I1206 03:46:16.695179 4980 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"2d3ca7fa81b93eca00b5d90c8b2685ff607a8156fe318cf99c5296fd65a5be75"} Dec 06 03:46:16 crc kubenswrapper[4980]: I1206 03:46:16.695341 4980 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"33fa9bd73b9af5df6de4e7ed019a512334f465b830a3f4396cde14989658bf55"} Dec 06 03:46:16 crc kubenswrapper[4980]: I1206 03:46:16.695463 4980 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"fe261f4322b85702181d7473ab2a9a4d2d4b869d586c3091f811b12a81a63aae"} Dec 06 03:46:16 crc kubenswrapper[4980]: I1206 03:46:16.695795 4980 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"2d620685574db7594132ae24a91ec0a93dce20a9a289540deebb39650b8a1816"} Dec 06 03:46:16 crc kubenswrapper[4980]: I1206 03:46:16.695917 4980 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"c67ec8f9c1764fcdf215d52822c2d4f0437dc7f056f05d708767121f6c1f3515"} Dec 06 03:46:16 crc kubenswrapper[4980]: I1206 03:46:16.696024 4980 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"6342903b318317ee8d84c2f5d467cb25c811a789be0d6ab0e0cba37b5611e16d"} Dec 06 03:46:16 crc kubenswrapper[4980]: I1206 03:46:16.696151 4980 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"768b8e1fe1a00a8467d744da5ad3f61006b0a3612c96f8e4d0ee6fcb93195f9f"} Dec 06 03:46:16 crc kubenswrapper[4980]: I1206 03:46:16.696273 4980 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"f35beae5339eeaf26494ed678e3c1bd02349eae2a4b4615325b17923f3b653b9"} Dec 06 03:46:16 crc kubenswrapper[4980]: I1206 03:46:16.696388 4980 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"a940ec6ee523313728f7e8ae934b7e0b0cff82771e3e64723309db1fae013c3a"} Dec 06 03:46:16 crc kubenswrapper[4980]: I1206 03:46:16.696494 4980 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"d876df04b5ab0ba07a5cbcafbe4648403c5014c59eb7d40adee2657c3a255473"} Dec 06 03:46:16 crc kubenswrapper[4980]: I1206 03:46:16.696661 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5w4fr" event={"ID":"5af0e768-3c9c-4401-ab56-57516cd3170e","Type":"ContainerDied","Data":"f35beae5339eeaf26494ed678e3c1bd02349eae2a4b4615325b17923f3b653b9"} Dec 06 03:46:16 crc kubenswrapper[4980]: I1206 03:46:16.696779 4980 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"2d3ca7fa81b93eca00b5d90c8b2685ff607a8156fe318cf99c5296fd65a5be75"} Dec 06 03:46:16 crc kubenswrapper[4980]: I1206 03:46:16.697059 4980 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"33fa9bd73b9af5df6de4e7ed019a512334f465b830a3f4396cde14989658bf55"} Dec 06 03:46:16 crc kubenswrapper[4980]: I1206 03:46:16.697177 4980 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"fe261f4322b85702181d7473ab2a9a4d2d4b869d586c3091f811b12a81a63aae"} Dec 06 03:46:16 crc kubenswrapper[4980]: I1206 03:46:16.697293 4980 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"2d620685574db7594132ae24a91ec0a93dce20a9a289540deebb39650b8a1816"} Dec 06 03:46:16 crc kubenswrapper[4980]: I1206 03:46:16.697402 4980 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"c67ec8f9c1764fcdf215d52822c2d4f0437dc7f056f05d708767121f6c1f3515"} Dec 06 03:46:16 crc kubenswrapper[4980]: I1206 03:46:16.697536 4980 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"6342903b318317ee8d84c2f5d467cb25c811a789be0d6ab0e0cba37b5611e16d"} Dec 06 03:46:16 crc kubenswrapper[4980]: I1206 03:46:16.697768 4980 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"768b8e1fe1a00a8467d744da5ad3f61006b0a3612c96f8e4d0ee6fcb93195f9f"} Dec 06 03:46:16 crc kubenswrapper[4980]: I1206 03:46:16.698276 4980 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"f35beae5339eeaf26494ed678e3c1bd02349eae2a4b4615325b17923f3b653b9"} Dec 06 03:46:16 crc kubenswrapper[4980]: I1206 03:46:16.698474 4980 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"a940ec6ee523313728f7e8ae934b7e0b0cff82771e3e64723309db1fae013c3a"} Dec 06 03:46:16 crc kubenswrapper[4980]: I1206 03:46:16.698647 4980 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"d876df04b5ab0ba07a5cbcafbe4648403c5014c59eb7d40adee2657c3a255473"} Dec 06 03:46:16 crc kubenswrapper[4980]: I1206 03:46:16.698787 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5w4fr" event={"ID":"5af0e768-3c9c-4401-ab56-57516cd3170e","Type":"ContainerDied","Data":"a940ec6ee523313728f7e8ae934b7e0b0cff82771e3e64723309db1fae013c3a"} Dec 06 03:46:16 crc kubenswrapper[4980]: I1206 03:46:16.698904 4980 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"2d3ca7fa81b93eca00b5d90c8b2685ff607a8156fe318cf99c5296fd65a5be75"} Dec 06 03:46:16 crc kubenswrapper[4980]: I1206 03:46:16.699029 4980 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"33fa9bd73b9af5df6de4e7ed019a512334f465b830a3f4396cde14989658bf55"} Dec 06 03:46:16 crc kubenswrapper[4980]: I1206 03:46:16.699137 4980 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"fe261f4322b85702181d7473ab2a9a4d2d4b869d586c3091f811b12a81a63aae"} Dec 06 03:46:16 crc kubenswrapper[4980]: I1206 03:46:16.699242 4980 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"2d620685574db7594132ae24a91ec0a93dce20a9a289540deebb39650b8a1816"} Dec 06 03:46:16 crc kubenswrapper[4980]: I1206 03:46:16.699346 4980 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"c67ec8f9c1764fcdf215d52822c2d4f0437dc7f056f05d708767121f6c1f3515"} Dec 06 03:46:16 crc kubenswrapper[4980]: I1206 03:46:16.699447 4980 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"6342903b318317ee8d84c2f5d467cb25c811a789be0d6ab0e0cba37b5611e16d"} Dec 06 03:46:16 crc kubenswrapper[4980]: I1206 03:46:16.699617 4980 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"768b8e1fe1a00a8467d744da5ad3f61006b0a3612c96f8e4d0ee6fcb93195f9f"} Dec 06 03:46:16 crc kubenswrapper[4980]: I1206 03:46:16.699803 4980 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"f35beae5339eeaf26494ed678e3c1bd02349eae2a4b4615325b17923f3b653b9"} Dec 06 03:46:16 crc kubenswrapper[4980]: I1206 03:46:16.699909 4980 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"a940ec6ee523313728f7e8ae934b7e0b0cff82771e3e64723309db1fae013c3a"} Dec 06 03:46:16 crc kubenswrapper[4980]: I1206 03:46:16.700039 4980 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"d876df04b5ab0ba07a5cbcafbe4648403c5014c59eb7d40adee2657c3a255473"} Dec 06 03:46:16 crc kubenswrapper[4980]: I1206 03:46:16.700173 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5w4fr" event={"ID":"5af0e768-3c9c-4401-ab56-57516cd3170e","Type":"ContainerDied","Data":"c6155f097866dc8f452b08fe7fb793ca7d6cf707ecfbb096967ad970e78f6cb9"} Dec 06 03:46:16 crc kubenswrapper[4980]: I1206 03:46:16.700287 4980 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"2d3ca7fa81b93eca00b5d90c8b2685ff607a8156fe318cf99c5296fd65a5be75"} Dec 06 03:46:16 crc kubenswrapper[4980]: I1206 03:46:16.700417 4980 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"33fa9bd73b9af5df6de4e7ed019a512334f465b830a3f4396cde14989658bf55"} Dec 06 03:46:16 crc kubenswrapper[4980]: I1206 03:46:16.700558 4980 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"fe261f4322b85702181d7473ab2a9a4d2d4b869d586c3091f811b12a81a63aae"} Dec 06 03:46:16 crc kubenswrapper[4980]: I1206 03:46:16.700662 4980 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"2d620685574db7594132ae24a91ec0a93dce20a9a289540deebb39650b8a1816"} Dec 06 03:46:16 crc kubenswrapper[4980]: I1206 03:46:16.700776 4980 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"c67ec8f9c1764fcdf215d52822c2d4f0437dc7f056f05d708767121f6c1f3515"} Dec 06 03:46:16 crc kubenswrapper[4980]: I1206 03:46:16.700911 4980 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"6342903b318317ee8d84c2f5d467cb25c811a789be0d6ab0e0cba37b5611e16d"} Dec 06 03:46:16 crc kubenswrapper[4980]: I1206 03:46:16.701023 4980 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"768b8e1fe1a00a8467d744da5ad3f61006b0a3612c96f8e4d0ee6fcb93195f9f"} Dec 06 03:46:16 crc kubenswrapper[4980]: I1206 03:46:16.701119 4980 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"f35beae5339eeaf26494ed678e3c1bd02349eae2a4b4615325b17923f3b653b9"} Dec 06 03:46:16 crc kubenswrapper[4980]: I1206 03:46:16.701254 4980 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"a940ec6ee523313728f7e8ae934b7e0b0cff82771e3e64723309db1fae013c3a"} Dec 06 03:46:16 crc kubenswrapper[4980]: I1206 03:46:16.701357 4980 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"d876df04b5ab0ba07a5cbcafbe4648403c5014c59eb7d40adee2657c3a255473"} Dec 06 03:46:16 crc kubenswrapper[4980]: I1206 03:46:16.721339 4980 scope.go:117] "RemoveContainer" containerID="2d3ca7fa81b93eca00b5d90c8b2685ff607a8156fe318cf99c5296fd65a5be75" Dec 06 03:46:16 crc kubenswrapper[4980]: I1206 03:46:16.749761 4980 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-5w4fr"] Dec 06 03:46:16 crc kubenswrapper[4980]: I1206 03:46:16.750816 4980 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-5w4fr"] Dec 06 03:46:16 crc kubenswrapper[4980]: I1206 03:46:16.756558 4980 scope.go:117] "RemoveContainer" containerID="33fa9bd73b9af5df6de4e7ed019a512334f465b830a3f4396cde14989658bf55" Dec 06 03:46:16 crc kubenswrapper[4980]: I1206 03:46:16.774964 4980 scope.go:117] "RemoveContainer" containerID="fe261f4322b85702181d7473ab2a9a4d2d4b869d586c3091f811b12a81a63aae" Dec 06 03:46:16 crc kubenswrapper[4980]: I1206 03:46:16.790717 4980 scope.go:117] "RemoveContainer" containerID="2d620685574db7594132ae24a91ec0a93dce20a9a289540deebb39650b8a1816" Dec 06 03:46:16 crc kubenswrapper[4980]: I1206 03:46:16.809535 4980 scope.go:117] "RemoveContainer" containerID="c67ec8f9c1764fcdf215d52822c2d4f0437dc7f056f05d708767121f6c1f3515" Dec 06 03:46:16 crc kubenswrapper[4980]: I1206 03:46:16.829073 4980 scope.go:117] "RemoveContainer" containerID="6342903b318317ee8d84c2f5d467cb25c811a789be0d6ab0e0cba37b5611e16d" Dec 06 03:46:16 crc kubenswrapper[4980]: I1206 03:46:16.848643 4980 scope.go:117] "RemoveContainer" containerID="768b8e1fe1a00a8467d744da5ad3f61006b0a3612c96f8e4d0ee6fcb93195f9f" Dec 06 03:46:16 crc kubenswrapper[4980]: I1206 03:46:16.864891 4980 scope.go:117] "RemoveContainer" containerID="f35beae5339eeaf26494ed678e3c1bd02349eae2a4b4615325b17923f3b653b9" Dec 06 03:46:16 crc kubenswrapper[4980]: I1206 03:46:16.876390 4980 scope.go:117] "RemoveContainer" containerID="a940ec6ee523313728f7e8ae934b7e0b0cff82771e3e64723309db1fae013c3a" Dec 06 03:46:16 crc kubenswrapper[4980]: I1206 03:46:16.911934 4980 scope.go:117] "RemoveContainer" containerID="d876df04b5ab0ba07a5cbcafbe4648403c5014c59eb7d40adee2657c3a255473" Dec 06 03:46:16 crc kubenswrapper[4980]: I1206 03:46:16.928374 4980 scope.go:117] "RemoveContainer" containerID="2d3ca7fa81b93eca00b5d90c8b2685ff607a8156fe318cf99c5296fd65a5be75" Dec 06 03:46:16 crc kubenswrapper[4980]: E1206 03:46:16.928933 4980 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2d3ca7fa81b93eca00b5d90c8b2685ff607a8156fe318cf99c5296fd65a5be75\": container with ID starting with 2d3ca7fa81b93eca00b5d90c8b2685ff607a8156fe318cf99c5296fd65a5be75 not found: ID does not exist" containerID="2d3ca7fa81b93eca00b5d90c8b2685ff607a8156fe318cf99c5296fd65a5be75" Dec 06 03:46:16 crc kubenswrapper[4980]: I1206 03:46:16.928981 4980 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2d3ca7fa81b93eca00b5d90c8b2685ff607a8156fe318cf99c5296fd65a5be75"} err="failed to get container status \"2d3ca7fa81b93eca00b5d90c8b2685ff607a8156fe318cf99c5296fd65a5be75\": rpc error: code = NotFound desc = could not find container \"2d3ca7fa81b93eca00b5d90c8b2685ff607a8156fe318cf99c5296fd65a5be75\": container with ID starting with 2d3ca7fa81b93eca00b5d90c8b2685ff607a8156fe318cf99c5296fd65a5be75 not found: ID does not exist" Dec 06 03:46:16 crc kubenswrapper[4980]: I1206 03:46:16.929011 4980 scope.go:117] "RemoveContainer" containerID="33fa9bd73b9af5df6de4e7ed019a512334f465b830a3f4396cde14989658bf55" Dec 06 03:46:16 crc kubenswrapper[4980]: E1206 03:46:16.929294 4980 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"33fa9bd73b9af5df6de4e7ed019a512334f465b830a3f4396cde14989658bf55\": container with ID starting with 33fa9bd73b9af5df6de4e7ed019a512334f465b830a3f4396cde14989658bf55 not found: ID does not exist" containerID="33fa9bd73b9af5df6de4e7ed019a512334f465b830a3f4396cde14989658bf55" Dec 06 03:46:16 crc kubenswrapper[4980]: I1206 03:46:16.929316 4980 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"33fa9bd73b9af5df6de4e7ed019a512334f465b830a3f4396cde14989658bf55"} err="failed to get container status \"33fa9bd73b9af5df6de4e7ed019a512334f465b830a3f4396cde14989658bf55\": rpc error: code = NotFound desc = could not find container \"33fa9bd73b9af5df6de4e7ed019a512334f465b830a3f4396cde14989658bf55\": container with ID starting with 33fa9bd73b9af5df6de4e7ed019a512334f465b830a3f4396cde14989658bf55 not found: ID does not exist" Dec 06 03:46:16 crc kubenswrapper[4980]: I1206 03:46:16.929329 4980 scope.go:117] "RemoveContainer" containerID="fe261f4322b85702181d7473ab2a9a4d2d4b869d586c3091f811b12a81a63aae" Dec 06 03:46:16 crc kubenswrapper[4980]: E1206 03:46:16.929799 4980 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fe261f4322b85702181d7473ab2a9a4d2d4b869d586c3091f811b12a81a63aae\": container with ID starting with fe261f4322b85702181d7473ab2a9a4d2d4b869d586c3091f811b12a81a63aae not found: ID does not exist" containerID="fe261f4322b85702181d7473ab2a9a4d2d4b869d586c3091f811b12a81a63aae" Dec 06 03:46:16 crc kubenswrapper[4980]: I1206 03:46:16.929822 4980 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fe261f4322b85702181d7473ab2a9a4d2d4b869d586c3091f811b12a81a63aae"} err="failed to get container status \"fe261f4322b85702181d7473ab2a9a4d2d4b869d586c3091f811b12a81a63aae\": rpc error: code = NotFound desc = could not find container \"fe261f4322b85702181d7473ab2a9a4d2d4b869d586c3091f811b12a81a63aae\": container with ID starting with fe261f4322b85702181d7473ab2a9a4d2d4b869d586c3091f811b12a81a63aae not found: ID does not exist" Dec 06 03:46:16 crc kubenswrapper[4980]: I1206 03:46:16.929835 4980 scope.go:117] "RemoveContainer" containerID="2d620685574db7594132ae24a91ec0a93dce20a9a289540deebb39650b8a1816" Dec 06 03:46:16 crc kubenswrapper[4980]: E1206 03:46:16.930217 4980 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2d620685574db7594132ae24a91ec0a93dce20a9a289540deebb39650b8a1816\": container with ID starting with 2d620685574db7594132ae24a91ec0a93dce20a9a289540deebb39650b8a1816 not found: ID does not exist" containerID="2d620685574db7594132ae24a91ec0a93dce20a9a289540deebb39650b8a1816" Dec 06 03:46:16 crc kubenswrapper[4980]: I1206 03:46:16.930252 4980 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2d620685574db7594132ae24a91ec0a93dce20a9a289540deebb39650b8a1816"} err="failed to get container status \"2d620685574db7594132ae24a91ec0a93dce20a9a289540deebb39650b8a1816\": rpc error: code = NotFound desc = could not find container \"2d620685574db7594132ae24a91ec0a93dce20a9a289540deebb39650b8a1816\": container with ID starting with 2d620685574db7594132ae24a91ec0a93dce20a9a289540deebb39650b8a1816 not found: ID does not exist" Dec 06 03:46:16 crc kubenswrapper[4980]: I1206 03:46:16.930265 4980 scope.go:117] "RemoveContainer" containerID="c67ec8f9c1764fcdf215d52822c2d4f0437dc7f056f05d708767121f6c1f3515" Dec 06 03:46:16 crc kubenswrapper[4980]: E1206 03:46:16.930458 4980 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c67ec8f9c1764fcdf215d52822c2d4f0437dc7f056f05d708767121f6c1f3515\": container with ID starting with c67ec8f9c1764fcdf215d52822c2d4f0437dc7f056f05d708767121f6c1f3515 not found: ID does not exist" containerID="c67ec8f9c1764fcdf215d52822c2d4f0437dc7f056f05d708767121f6c1f3515" Dec 06 03:46:16 crc kubenswrapper[4980]: I1206 03:46:16.930480 4980 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c67ec8f9c1764fcdf215d52822c2d4f0437dc7f056f05d708767121f6c1f3515"} err="failed to get container status \"c67ec8f9c1764fcdf215d52822c2d4f0437dc7f056f05d708767121f6c1f3515\": rpc error: code = NotFound desc = could not find container \"c67ec8f9c1764fcdf215d52822c2d4f0437dc7f056f05d708767121f6c1f3515\": container with ID starting with c67ec8f9c1764fcdf215d52822c2d4f0437dc7f056f05d708767121f6c1f3515 not found: ID does not exist" Dec 06 03:46:16 crc kubenswrapper[4980]: I1206 03:46:16.930493 4980 scope.go:117] "RemoveContainer" containerID="6342903b318317ee8d84c2f5d467cb25c811a789be0d6ab0e0cba37b5611e16d" Dec 06 03:46:16 crc kubenswrapper[4980]: E1206 03:46:16.930816 4980 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6342903b318317ee8d84c2f5d467cb25c811a789be0d6ab0e0cba37b5611e16d\": container with ID starting with 6342903b318317ee8d84c2f5d467cb25c811a789be0d6ab0e0cba37b5611e16d not found: ID does not exist" containerID="6342903b318317ee8d84c2f5d467cb25c811a789be0d6ab0e0cba37b5611e16d" Dec 06 03:46:16 crc kubenswrapper[4980]: I1206 03:46:16.930837 4980 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6342903b318317ee8d84c2f5d467cb25c811a789be0d6ab0e0cba37b5611e16d"} err="failed to get container status \"6342903b318317ee8d84c2f5d467cb25c811a789be0d6ab0e0cba37b5611e16d\": rpc error: code = NotFound desc = could not find container \"6342903b318317ee8d84c2f5d467cb25c811a789be0d6ab0e0cba37b5611e16d\": container with ID starting with 6342903b318317ee8d84c2f5d467cb25c811a789be0d6ab0e0cba37b5611e16d not found: ID does not exist" Dec 06 03:46:16 crc kubenswrapper[4980]: I1206 03:46:16.930856 4980 scope.go:117] "RemoveContainer" containerID="768b8e1fe1a00a8467d744da5ad3f61006b0a3612c96f8e4d0ee6fcb93195f9f" Dec 06 03:46:16 crc kubenswrapper[4980]: E1206 03:46:16.931031 4980 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"768b8e1fe1a00a8467d744da5ad3f61006b0a3612c96f8e4d0ee6fcb93195f9f\": container with ID starting with 768b8e1fe1a00a8467d744da5ad3f61006b0a3612c96f8e4d0ee6fcb93195f9f not found: ID does not exist" containerID="768b8e1fe1a00a8467d744da5ad3f61006b0a3612c96f8e4d0ee6fcb93195f9f" Dec 06 03:46:16 crc kubenswrapper[4980]: I1206 03:46:16.931061 4980 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"768b8e1fe1a00a8467d744da5ad3f61006b0a3612c96f8e4d0ee6fcb93195f9f"} err="failed to get container status \"768b8e1fe1a00a8467d744da5ad3f61006b0a3612c96f8e4d0ee6fcb93195f9f\": rpc error: code = NotFound desc = could not find container \"768b8e1fe1a00a8467d744da5ad3f61006b0a3612c96f8e4d0ee6fcb93195f9f\": container with ID starting with 768b8e1fe1a00a8467d744da5ad3f61006b0a3612c96f8e4d0ee6fcb93195f9f not found: ID does not exist" Dec 06 03:46:16 crc kubenswrapper[4980]: I1206 03:46:16.931079 4980 scope.go:117] "RemoveContainer" containerID="f35beae5339eeaf26494ed678e3c1bd02349eae2a4b4615325b17923f3b653b9" Dec 06 03:46:16 crc kubenswrapper[4980]: E1206 03:46:16.931437 4980 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f35beae5339eeaf26494ed678e3c1bd02349eae2a4b4615325b17923f3b653b9\": container with ID starting with f35beae5339eeaf26494ed678e3c1bd02349eae2a4b4615325b17923f3b653b9 not found: ID does not exist" containerID="f35beae5339eeaf26494ed678e3c1bd02349eae2a4b4615325b17923f3b653b9" Dec 06 03:46:16 crc kubenswrapper[4980]: I1206 03:46:16.931458 4980 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f35beae5339eeaf26494ed678e3c1bd02349eae2a4b4615325b17923f3b653b9"} err="failed to get container status \"f35beae5339eeaf26494ed678e3c1bd02349eae2a4b4615325b17923f3b653b9\": rpc error: code = NotFound desc = could not find container \"f35beae5339eeaf26494ed678e3c1bd02349eae2a4b4615325b17923f3b653b9\": container with ID starting with f35beae5339eeaf26494ed678e3c1bd02349eae2a4b4615325b17923f3b653b9 not found: ID does not exist" Dec 06 03:46:16 crc kubenswrapper[4980]: I1206 03:46:16.931470 4980 scope.go:117] "RemoveContainer" containerID="a940ec6ee523313728f7e8ae934b7e0b0cff82771e3e64723309db1fae013c3a" Dec 06 03:46:16 crc kubenswrapper[4980]: E1206 03:46:16.931723 4980 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a940ec6ee523313728f7e8ae934b7e0b0cff82771e3e64723309db1fae013c3a\": container with ID starting with a940ec6ee523313728f7e8ae934b7e0b0cff82771e3e64723309db1fae013c3a not found: ID does not exist" containerID="a940ec6ee523313728f7e8ae934b7e0b0cff82771e3e64723309db1fae013c3a" Dec 06 03:46:16 crc kubenswrapper[4980]: I1206 03:46:16.931742 4980 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a940ec6ee523313728f7e8ae934b7e0b0cff82771e3e64723309db1fae013c3a"} err="failed to get container status \"a940ec6ee523313728f7e8ae934b7e0b0cff82771e3e64723309db1fae013c3a\": rpc error: code = NotFound desc = could not find container \"a940ec6ee523313728f7e8ae934b7e0b0cff82771e3e64723309db1fae013c3a\": container with ID starting with a940ec6ee523313728f7e8ae934b7e0b0cff82771e3e64723309db1fae013c3a not found: ID does not exist" Dec 06 03:46:16 crc kubenswrapper[4980]: I1206 03:46:16.931757 4980 scope.go:117] "RemoveContainer" containerID="d876df04b5ab0ba07a5cbcafbe4648403c5014c59eb7d40adee2657c3a255473" Dec 06 03:46:16 crc kubenswrapper[4980]: E1206 03:46:16.932097 4980 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d876df04b5ab0ba07a5cbcafbe4648403c5014c59eb7d40adee2657c3a255473\": container with ID starting with d876df04b5ab0ba07a5cbcafbe4648403c5014c59eb7d40adee2657c3a255473 not found: ID does not exist" containerID="d876df04b5ab0ba07a5cbcafbe4648403c5014c59eb7d40adee2657c3a255473" Dec 06 03:46:16 crc kubenswrapper[4980]: I1206 03:46:16.932132 4980 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d876df04b5ab0ba07a5cbcafbe4648403c5014c59eb7d40adee2657c3a255473"} err="failed to get container status \"d876df04b5ab0ba07a5cbcafbe4648403c5014c59eb7d40adee2657c3a255473\": rpc error: code = NotFound desc = could not find container \"d876df04b5ab0ba07a5cbcafbe4648403c5014c59eb7d40adee2657c3a255473\": container with ID starting with d876df04b5ab0ba07a5cbcafbe4648403c5014c59eb7d40adee2657c3a255473 not found: ID does not exist" Dec 06 03:46:16 crc kubenswrapper[4980]: I1206 03:46:16.932149 4980 scope.go:117] "RemoveContainer" containerID="2d3ca7fa81b93eca00b5d90c8b2685ff607a8156fe318cf99c5296fd65a5be75" Dec 06 03:46:16 crc kubenswrapper[4980]: I1206 03:46:16.932498 4980 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2d3ca7fa81b93eca00b5d90c8b2685ff607a8156fe318cf99c5296fd65a5be75"} err="failed to get container status \"2d3ca7fa81b93eca00b5d90c8b2685ff607a8156fe318cf99c5296fd65a5be75\": rpc error: code = NotFound desc = could not find container \"2d3ca7fa81b93eca00b5d90c8b2685ff607a8156fe318cf99c5296fd65a5be75\": container with ID starting with 2d3ca7fa81b93eca00b5d90c8b2685ff607a8156fe318cf99c5296fd65a5be75 not found: ID does not exist" Dec 06 03:46:16 crc kubenswrapper[4980]: I1206 03:46:16.932529 4980 scope.go:117] "RemoveContainer" containerID="33fa9bd73b9af5df6de4e7ed019a512334f465b830a3f4396cde14989658bf55" Dec 06 03:46:16 crc kubenswrapper[4980]: I1206 03:46:16.932842 4980 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"33fa9bd73b9af5df6de4e7ed019a512334f465b830a3f4396cde14989658bf55"} err="failed to get container status \"33fa9bd73b9af5df6de4e7ed019a512334f465b830a3f4396cde14989658bf55\": rpc error: code = NotFound desc = could not find container \"33fa9bd73b9af5df6de4e7ed019a512334f465b830a3f4396cde14989658bf55\": container with ID starting with 33fa9bd73b9af5df6de4e7ed019a512334f465b830a3f4396cde14989658bf55 not found: ID does not exist" Dec 06 03:46:16 crc kubenswrapper[4980]: I1206 03:46:16.932864 4980 scope.go:117] "RemoveContainer" containerID="fe261f4322b85702181d7473ab2a9a4d2d4b869d586c3091f811b12a81a63aae" Dec 06 03:46:16 crc kubenswrapper[4980]: I1206 03:46:16.933137 4980 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fe261f4322b85702181d7473ab2a9a4d2d4b869d586c3091f811b12a81a63aae"} err="failed to get container status \"fe261f4322b85702181d7473ab2a9a4d2d4b869d586c3091f811b12a81a63aae\": rpc error: code = NotFound desc = could not find container \"fe261f4322b85702181d7473ab2a9a4d2d4b869d586c3091f811b12a81a63aae\": container with ID starting with fe261f4322b85702181d7473ab2a9a4d2d4b869d586c3091f811b12a81a63aae not found: ID does not exist" Dec 06 03:46:16 crc kubenswrapper[4980]: I1206 03:46:16.933167 4980 scope.go:117] "RemoveContainer" containerID="2d620685574db7594132ae24a91ec0a93dce20a9a289540deebb39650b8a1816" Dec 06 03:46:16 crc kubenswrapper[4980]: I1206 03:46:16.933564 4980 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2d620685574db7594132ae24a91ec0a93dce20a9a289540deebb39650b8a1816"} err="failed to get container status \"2d620685574db7594132ae24a91ec0a93dce20a9a289540deebb39650b8a1816\": rpc error: code = NotFound desc = could not find container \"2d620685574db7594132ae24a91ec0a93dce20a9a289540deebb39650b8a1816\": container with ID starting with 2d620685574db7594132ae24a91ec0a93dce20a9a289540deebb39650b8a1816 not found: ID does not exist" Dec 06 03:46:16 crc kubenswrapper[4980]: I1206 03:46:16.933583 4980 scope.go:117] "RemoveContainer" containerID="c67ec8f9c1764fcdf215d52822c2d4f0437dc7f056f05d708767121f6c1f3515" Dec 06 03:46:16 crc kubenswrapper[4980]: I1206 03:46:16.933866 4980 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c67ec8f9c1764fcdf215d52822c2d4f0437dc7f056f05d708767121f6c1f3515"} err="failed to get container status \"c67ec8f9c1764fcdf215d52822c2d4f0437dc7f056f05d708767121f6c1f3515\": rpc error: code = NotFound desc = could not find container \"c67ec8f9c1764fcdf215d52822c2d4f0437dc7f056f05d708767121f6c1f3515\": container with ID starting with c67ec8f9c1764fcdf215d52822c2d4f0437dc7f056f05d708767121f6c1f3515 not found: ID does not exist" Dec 06 03:46:16 crc kubenswrapper[4980]: I1206 03:46:16.933894 4980 scope.go:117] "RemoveContainer" containerID="6342903b318317ee8d84c2f5d467cb25c811a789be0d6ab0e0cba37b5611e16d" Dec 06 03:46:16 crc kubenswrapper[4980]: I1206 03:46:16.934071 4980 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6342903b318317ee8d84c2f5d467cb25c811a789be0d6ab0e0cba37b5611e16d"} err="failed to get container status \"6342903b318317ee8d84c2f5d467cb25c811a789be0d6ab0e0cba37b5611e16d\": rpc error: code = NotFound desc = could not find container \"6342903b318317ee8d84c2f5d467cb25c811a789be0d6ab0e0cba37b5611e16d\": container with ID starting with 6342903b318317ee8d84c2f5d467cb25c811a789be0d6ab0e0cba37b5611e16d not found: ID does not exist" Dec 06 03:46:16 crc kubenswrapper[4980]: I1206 03:46:16.934090 4980 scope.go:117] "RemoveContainer" containerID="768b8e1fe1a00a8467d744da5ad3f61006b0a3612c96f8e4d0ee6fcb93195f9f" Dec 06 03:46:16 crc kubenswrapper[4980]: I1206 03:46:16.934243 4980 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"768b8e1fe1a00a8467d744da5ad3f61006b0a3612c96f8e4d0ee6fcb93195f9f"} err="failed to get container status \"768b8e1fe1a00a8467d744da5ad3f61006b0a3612c96f8e4d0ee6fcb93195f9f\": rpc error: code = NotFound desc = could not find container \"768b8e1fe1a00a8467d744da5ad3f61006b0a3612c96f8e4d0ee6fcb93195f9f\": container with ID starting with 768b8e1fe1a00a8467d744da5ad3f61006b0a3612c96f8e4d0ee6fcb93195f9f not found: ID does not exist" Dec 06 03:46:16 crc kubenswrapper[4980]: I1206 03:46:16.934260 4980 scope.go:117] "RemoveContainer" containerID="f35beae5339eeaf26494ed678e3c1bd02349eae2a4b4615325b17923f3b653b9" Dec 06 03:46:16 crc kubenswrapper[4980]: I1206 03:46:16.934564 4980 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f35beae5339eeaf26494ed678e3c1bd02349eae2a4b4615325b17923f3b653b9"} err="failed to get container status \"f35beae5339eeaf26494ed678e3c1bd02349eae2a4b4615325b17923f3b653b9\": rpc error: code = NotFound desc = could not find container \"f35beae5339eeaf26494ed678e3c1bd02349eae2a4b4615325b17923f3b653b9\": container with ID starting with f35beae5339eeaf26494ed678e3c1bd02349eae2a4b4615325b17923f3b653b9 not found: ID does not exist" Dec 06 03:46:16 crc kubenswrapper[4980]: I1206 03:46:16.934583 4980 scope.go:117] "RemoveContainer" containerID="a940ec6ee523313728f7e8ae934b7e0b0cff82771e3e64723309db1fae013c3a" Dec 06 03:46:16 crc kubenswrapper[4980]: I1206 03:46:16.934772 4980 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a940ec6ee523313728f7e8ae934b7e0b0cff82771e3e64723309db1fae013c3a"} err="failed to get container status \"a940ec6ee523313728f7e8ae934b7e0b0cff82771e3e64723309db1fae013c3a\": rpc error: code = NotFound desc = could not find container \"a940ec6ee523313728f7e8ae934b7e0b0cff82771e3e64723309db1fae013c3a\": container with ID starting with a940ec6ee523313728f7e8ae934b7e0b0cff82771e3e64723309db1fae013c3a not found: ID does not exist" Dec 06 03:46:16 crc kubenswrapper[4980]: I1206 03:46:16.934785 4980 scope.go:117] "RemoveContainer" containerID="d876df04b5ab0ba07a5cbcafbe4648403c5014c59eb7d40adee2657c3a255473" Dec 06 03:46:16 crc kubenswrapper[4980]: I1206 03:46:16.934979 4980 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d876df04b5ab0ba07a5cbcafbe4648403c5014c59eb7d40adee2657c3a255473"} err="failed to get container status \"d876df04b5ab0ba07a5cbcafbe4648403c5014c59eb7d40adee2657c3a255473\": rpc error: code = NotFound desc = could not find container \"d876df04b5ab0ba07a5cbcafbe4648403c5014c59eb7d40adee2657c3a255473\": container with ID starting with d876df04b5ab0ba07a5cbcafbe4648403c5014c59eb7d40adee2657c3a255473 not found: ID does not exist" Dec 06 03:46:16 crc kubenswrapper[4980]: I1206 03:46:16.934993 4980 scope.go:117] "RemoveContainer" containerID="2d3ca7fa81b93eca00b5d90c8b2685ff607a8156fe318cf99c5296fd65a5be75" Dec 06 03:46:16 crc kubenswrapper[4980]: I1206 03:46:16.935220 4980 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2d3ca7fa81b93eca00b5d90c8b2685ff607a8156fe318cf99c5296fd65a5be75"} err="failed to get container status \"2d3ca7fa81b93eca00b5d90c8b2685ff607a8156fe318cf99c5296fd65a5be75\": rpc error: code = NotFound desc = could not find container \"2d3ca7fa81b93eca00b5d90c8b2685ff607a8156fe318cf99c5296fd65a5be75\": container with ID starting with 2d3ca7fa81b93eca00b5d90c8b2685ff607a8156fe318cf99c5296fd65a5be75 not found: ID does not exist" Dec 06 03:46:16 crc kubenswrapper[4980]: I1206 03:46:16.935234 4980 scope.go:117] "RemoveContainer" containerID="33fa9bd73b9af5df6de4e7ed019a512334f465b830a3f4396cde14989658bf55" Dec 06 03:46:16 crc kubenswrapper[4980]: I1206 03:46:16.935414 4980 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"33fa9bd73b9af5df6de4e7ed019a512334f465b830a3f4396cde14989658bf55"} err="failed to get container status \"33fa9bd73b9af5df6de4e7ed019a512334f465b830a3f4396cde14989658bf55\": rpc error: code = NotFound desc = could not find container \"33fa9bd73b9af5df6de4e7ed019a512334f465b830a3f4396cde14989658bf55\": container with ID starting with 33fa9bd73b9af5df6de4e7ed019a512334f465b830a3f4396cde14989658bf55 not found: ID does not exist" Dec 06 03:46:16 crc kubenswrapper[4980]: I1206 03:46:16.935433 4980 scope.go:117] "RemoveContainer" containerID="fe261f4322b85702181d7473ab2a9a4d2d4b869d586c3091f811b12a81a63aae" Dec 06 03:46:16 crc kubenswrapper[4980]: I1206 03:46:16.935652 4980 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fe261f4322b85702181d7473ab2a9a4d2d4b869d586c3091f811b12a81a63aae"} err="failed to get container status \"fe261f4322b85702181d7473ab2a9a4d2d4b869d586c3091f811b12a81a63aae\": rpc error: code = NotFound desc = could not find container \"fe261f4322b85702181d7473ab2a9a4d2d4b869d586c3091f811b12a81a63aae\": container with ID starting with fe261f4322b85702181d7473ab2a9a4d2d4b869d586c3091f811b12a81a63aae not found: ID does not exist" Dec 06 03:46:16 crc kubenswrapper[4980]: I1206 03:46:16.935669 4980 scope.go:117] "RemoveContainer" containerID="2d620685574db7594132ae24a91ec0a93dce20a9a289540deebb39650b8a1816" Dec 06 03:46:16 crc kubenswrapper[4980]: I1206 03:46:16.935970 4980 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2d620685574db7594132ae24a91ec0a93dce20a9a289540deebb39650b8a1816"} err="failed to get container status \"2d620685574db7594132ae24a91ec0a93dce20a9a289540deebb39650b8a1816\": rpc error: code = NotFound desc = could not find container \"2d620685574db7594132ae24a91ec0a93dce20a9a289540deebb39650b8a1816\": container with ID starting with 2d620685574db7594132ae24a91ec0a93dce20a9a289540deebb39650b8a1816 not found: ID does not exist" Dec 06 03:46:16 crc kubenswrapper[4980]: I1206 03:46:16.935999 4980 scope.go:117] "RemoveContainer" containerID="c67ec8f9c1764fcdf215d52822c2d4f0437dc7f056f05d708767121f6c1f3515" Dec 06 03:46:16 crc kubenswrapper[4980]: I1206 03:46:16.936283 4980 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c67ec8f9c1764fcdf215d52822c2d4f0437dc7f056f05d708767121f6c1f3515"} err="failed to get container status \"c67ec8f9c1764fcdf215d52822c2d4f0437dc7f056f05d708767121f6c1f3515\": rpc error: code = NotFound desc = could not find container \"c67ec8f9c1764fcdf215d52822c2d4f0437dc7f056f05d708767121f6c1f3515\": container with ID starting with c67ec8f9c1764fcdf215d52822c2d4f0437dc7f056f05d708767121f6c1f3515 not found: ID does not exist" Dec 06 03:46:16 crc kubenswrapper[4980]: I1206 03:46:16.936300 4980 scope.go:117] "RemoveContainer" containerID="6342903b318317ee8d84c2f5d467cb25c811a789be0d6ab0e0cba37b5611e16d" Dec 06 03:46:16 crc kubenswrapper[4980]: I1206 03:46:16.936757 4980 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6342903b318317ee8d84c2f5d467cb25c811a789be0d6ab0e0cba37b5611e16d"} err="failed to get container status \"6342903b318317ee8d84c2f5d467cb25c811a789be0d6ab0e0cba37b5611e16d\": rpc error: code = NotFound desc = could not find container \"6342903b318317ee8d84c2f5d467cb25c811a789be0d6ab0e0cba37b5611e16d\": container with ID starting with 6342903b318317ee8d84c2f5d467cb25c811a789be0d6ab0e0cba37b5611e16d not found: ID does not exist" Dec 06 03:46:16 crc kubenswrapper[4980]: I1206 03:46:16.936778 4980 scope.go:117] "RemoveContainer" containerID="768b8e1fe1a00a8467d744da5ad3f61006b0a3612c96f8e4d0ee6fcb93195f9f" Dec 06 03:46:16 crc kubenswrapper[4980]: I1206 03:46:16.937096 4980 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"768b8e1fe1a00a8467d744da5ad3f61006b0a3612c96f8e4d0ee6fcb93195f9f"} err="failed to get container status \"768b8e1fe1a00a8467d744da5ad3f61006b0a3612c96f8e4d0ee6fcb93195f9f\": rpc error: code = NotFound desc = could not find container \"768b8e1fe1a00a8467d744da5ad3f61006b0a3612c96f8e4d0ee6fcb93195f9f\": container with ID starting with 768b8e1fe1a00a8467d744da5ad3f61006b0a3612c96f8e4d0ee6fcb93195f9f not found: ID does not exist" Dec 06 03:46:16 crc kubenswrapper[4980]: I1206 03:46:16.937115 4980 scope.go:117] "RemoveContainer" containerID="f35beae5339eeaf26494ed678e3c1bd02349eae2a4b4615325b17923f3b653b9" Dec 06 03:46:16 crc kubenswrapper[4980]: I1206 03:46:16.937280 4980 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f35beae5339eeaf26494ed678e3c1bd02349eae2a4b4615325b17923f3b653b9"} err="failed to get container status \"f35beae5339eeaf26494ed678e3c1bd02349eae2a4b4615325b17923f3b653b9\": rpc error: code = NotFound desc = could not find container \"f35beae5339eeaf26494ed678e3c1bd02349eae2a4b4615325b17923f3b653b9\": container with ID starting with f35beae5339eeaf26494ed678e3c1bd02349eae2a4b4615325b17923f3b653b9 not found: ID does not exist" Dec 06 03:46:16 crc kubenswrapper[4980]: I1206 03:46:16.937293 4980 scope.go:117] "RemoveContainer" containerID="a940ec6ee523313728f7e8ae934b7e0b0cff82771e3e64723309db1fae013c3a" Dec 06 03:46:16 crc kubenswrapper[4980]: I1206 03:46:16.937548 4980 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a940ec6ee523313728f7e8ae934b7e0b0cff82771e3e64723309db1fae013c3a"} err="failed to get container status \"a940ec6ee523313728f7e8ae934b7e0b0cff82771e3e64723309db1fae013c3a\": rpc error: code = NotFound desc = could not find container \"a940ec6ee523313728f7e8ae934b7e0b0cff82771e3e64723309db1fae013c3a\": container with ID starting with a940ec6ee523313728f7e8ae934b7e0b0cff82771e3e64723309db1fae013c3a not found: ID does not exist" Dec 06 03:46:16 crc kubenswrapper[4980]: I1206 03:46:16.937567 4980 scope.go:117] "RemoveContainer" containerID="d876df04b5ab0ba07a5cbcafbe4648403c5014c59eb7d40adee2657c3a255473" Dec 06 03:46:16 crc kubenswrapper[4980]: I1206 03:46:16.937865 4980 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d876df04b5ab0ba07a5cbcafbe4648403c5014c59eb7d40adee2657c3a255473"} err="failed to get container status \"d876df04b5ab0ba07a5cbcafbe4648403c5014c59eb7d40adee2657c3a255473\": rpc error: code = NotFound desc = could not find container \"d876df04b5ab0ba07a5cbcafbe4648403c5014c59eb7d40adee2657c3a255473\": container with ID starting with d876df04b5ab0ba07a5cbcafbe4648403c5014c59eb7d40adee2657c3a255473 not found: ID does not exist" Dec 06 03:46:16 crc kubenswrapper[4980]: I1206 03:46:16.937894 4980 scope.go:117] "RemoveContainer" containerID="2d3ca7fa81b93eca00b5d90c8b2685ff607a8156fe318cf99c5296fd65a5be75" Dec 06 03:46:16 crc kubenswrapper[4980]: I1206 03:46:16.938091 4980 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2d3ca7fa81b93eca00b5d90c8b2685ff607a8156fe318cf99c5296fd65a5be75"} err="failed to get container status \"2d3ca7fa81b93eca00b5d90c8b2685ff607a8156fe318cf99c5296fd65a5be75\": rpc error: code = NotFound desc = could not find container \"2d3ca7fa81b93eca00b5d90c8b2685ff607a8156fe318cf99c5296fd65a5be75\": container with ID starting with 2d3ca7fa81b93eca00b5d90c8b2685ff607a8156fe318cf99c5296fd65a5be75 not found: ID does not exist" Dec 06 03:46:16 crc kubenswrapper[4980]: I1206 03:46:16.938103 4980 scope.go:117] "RemoveContainer" containerID="33fa9bd73b9af5df6de4e7ed019a512334f465b830a3f4396cde14989658bf55" Dec 06 03:46:16 crc kubenswrapper[4980]: I1206 03:46:16.938338 4980 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"33fa9bd73b9af5df6de4e7ed019a512334f465b830a3f4396cde14989658bf55"} err="failed to get container status \"33fa9bd73b9af5df6de4e7ed019a512334f465b830a3f4396cde14989658bf55\": rpc error: code = NotFound desc = could not find container \"33fa9bd73b9af5df6de4e7ed019a512334f465b830a3f4396cde14989658bf55\": container with ID starting with 33fa9bd73b9af5df6de4e7ed019a512334f465b830a3f4396cde14989658bf55 not found: ID does not exist" Dec 06 03:46:16 crc kubenswrapper[4980]: I1206 03:46:16.938354 4980 scope.go:117] "RemoveContainer" containerID="fe261f4322b85702181d7473ab2a9a4d2d4b869d586c3091f811b12a81a63aae" Dec 06 03:46:16 crc kubenswrapper[4980]: I1206 03:46:16.938674 4980 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fe261f4322b85702181d7473ab2a9a4d2d4b869d586c3091f811b12a81a63aae"} err="failed to get container status \"fe261f4322b85702181d7473ab2a9a4d2d4b869d586c3091f811b12a81a63aae\": rpc error: code = NotFound desc = could not find container \"fe261f4322b85702181d7473ab2a9a4d2d4b869d586c3091f811b12a81a63aae\": container with ID starting with fe261f4322b85702181d7473ab2a9a4d2d4b869d586c3091f811b12a81a63aae not found: ID does not exist" Dec 06 03:46:16 crc kubenswrapper[4980]: I1206 03:46:16.938722 4980 scope.go:117] "RemoveContainer" containerID="2d620685574db7594132ae24a91ec0a93dce20a9a289540deebb39650b8a1816" Dec 06 03:46:16 crc kubenswrapper[4980]: I1206 03:46:16.939057 4980 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2d620685574db7594132ae24a91ec0a93dce20a9a289540deebb39650b8a1816"} err="failed to get container status \"2d620685574db7594132ae24a91ec0a93dce20a9a289540deebb39650b8a1816\": rpc error: code = NotFound desc = could not find container \"2d620685574db7594132ae24a91ec0a93dce20a9a289540deebb39650b8a1816\": container with ID starting with 2d620685574db7594132ae24a91ec0a93dce20a9a289540deebb39650b8a1816 not found: ID does not exist" Dec 06 03:46:16 crc kubenswrapper[4980]: I1206 03:46:16.939113 4980 scope.go:117] "RemoveContainer" containerID="c67ec8f9c1764fcdf215d52822c2d4f0437dc7f056f05d708767121f6c1f3515" Dec 06 03:46:16 crc kubenswrapper[4980]: I1206 03:46:16.939391 4980 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c67ec8f9c1764fcdf215d52822c2d4f0437dc7f056f05d708767121f6c1f3515"} err="failed to get container status \"c67ec8f9c1764fcdf215d52822c2d4f0437dc7f056f05d708767121f6c1f3515\": rpc error: code = NotFound desc = could not find container \"c67ec8f9c1764fcdf215d52822c2d4f0437dc7f056f05d708767121f6c1f3515\": container with ID starting with c67ec8f9c1764fcdf215d52822c2d4f0437dc7f056f05d708767121f6c1f3515 not found: ID does not exist" Dec 06 03:46:16 crc kubenswrapper[4980]: I1206 03:46:16.939410 4980 scope.go:117] "RemoveContainer" containerID="6342903b318317ee8d84c2f5d467cb25c811a789be0d6ab0e0cba37b5611e16d" Dec 06 03:46:16 crc kubenswrapper[4980]: I1206 03:46:16.939700 4980 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6342903b318317ee8d84c2f5d467cb25c811a789be0d6ab0e0cba37b5611e16d"} err="failed to get container status \"6342903b318317ee8d84c2f5d467cb25c811a789be0d6ab0e0cba37b5611e16d\": rpc error: code = NotFound desc = could not find container \"6342903b318317ee8d84c2f5d467cb25c811a789be0d6ab0e0cba37b5611e16d\": container with ID starting with 6342903b318317ee8d84c2f5d467cb25c811a789be0d6ab0e0cba37b5611e16d not found: ID does not exist" Dec 06 03:46:16 crc kubenswrapper[4980]: I1206 03:46:16.939716 4980 scope.go:117] "RemoveContainer" containerID="768b8e1fe1a00a8467d744da5ad3f61006b0a3612c96f8e4d0ee6fcb93195f9f" Dec 06 03:46:16 crc kubenswrapper[4980]: I1206 03:46:16.939985 4980 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"768b8e1fe1a00a8467d744da5ad3f61006b0a3612c96f8e4d0ee6fcb93195f9f"} err="failed to get container status \"768b8e1fe1a00a8467d744da5ad3f61006b0a3612c96f8e4d0ee6fcb93195f9f\": rpc error: code = NotFound desc = could not find container \"768b8e1fe1a00a8467d744da5ad3f61006b0a3612c96f8e4d0ee6fcb93195f9f\": container with ID starting with 768b8e1fe1a00a8467d744da5ad3f61006b0a3612c96f8e4d0ee6fcb93195f9f not found: ID does not exist" Dec 06 03:46:16 crc kubenswrapper[4980]: I1206 03:46:16.940006 4980 scope.go:117] "RemoveContainer" containerID="f35beae5339eeaf26494ed678e3c1bd02349eae2a4b4615325b17923f3b653b9" Dec 06 03:46:16 crc kubenswrapper[4980]: I1206 03:46:16.940275 4980 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f35beae5339eeaf26494ed678e3c1bd02349eae2a4b4615325b17923f3b653b9"} err="failed to get container status \"f35beae5339eeaf26494ed678e3c1bd02349eae2a4b4615325b17923f3b653b9\": rpc error: code = NotFound desc = could not find container \"f35beae5339eeaf26494ed678e3c1bd02349eae2a4b4615325b17923f3b653b9\": container with ID starting with f35beae5339eeaf26494ed678e3c1bd02349eae2a4b4615325b17923f3b653b9 not found: ID does not exist" Dec 06 03:46:16 crc kubenswrapper[4980]: I1206 03:46:16.940300 4980 scope.go:117] "RemoveContainer" containerID="a940ec6ee523313728f7e8ae934b7e0b0cff82771e3e64723309db1fae013c3a" Dec 06 03:46:16 crc kubenswrapper[4980]: I1206 03:46:16.940714 4980 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a940ec6ee523313728f7e8ae934b7e0b0cff82771e3e64723309db1fae013c3a"} err="failed to get container status \"a940ec6ee523313728f7e8ae934b7e0b0cff82771e3e64723309db1fae013c3a\": rpc error: code = NotFound desc = could not find container \"a940ec6ee523313728f7e8ae934b7e0b0cff82771e3e64723309db1fae013c3a\": container with ID starting with a940ec6ee523313728f7e8ae934b7e0b0cff82771e3e64723309db1fae013c3a not found: ID does not exist" Dec 06 03:46:16 crc kubenswrapper[4980]: I1206 03:46:16.940758 4980 scope.go:117] "RemoveContainer" containerID="d876df04b5ab0ba07a5cbcafbe4648403c5014c59eb7d40adee2657c3a255473" Dec 06 03:46:16 crc kubenswrapper[4980]: I1206 03:46:16.941000 4980 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d876df04b5ab0ba07a5cbcafbe4648403c5014c59eb7d40adee2657c3a255473"} err="failed to get container status \"d876df04b5ab0ba07a5cbcafbe4648403c5014c59eb7d40adee2657c3a255473\": rpc error: code = NotFound desc = could not find container \"d876df04b5ab0ba07a5cbcafbe4648403c5014c59eb7d40adee2657c3a255473\": container with ID starting with d876df04b5ab0ba07a5cbcafbe4648403c5014c59eb7d40adee2657c3a255473 not found: ID does not exist" Dec 06 03:46:17 crc kubenswrapper[4980]: I1206 03:46:17.197137 4980 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5af0e768-3c9c-4401-ab56-57516cd3170e" path="/var/lib/kubelet/pods/5af0e768-3c9c-4401-ab56-57516cd3170e/volumes" Dec 06 03:46:17 crc kubenswrapper[4980]: I1206 03:46:17.704822 4980 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-kq4z8_f96c1f6c-55e8-4d86-bff3-5381581b7022/kube-multus/2.log" Dec 06 03:46:17 crc kubenswrapper[4980]: I1206 03:46:17.704973 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-kq4z8" event={"ID":"f96c1f6c-55e8-4d86-bff3-5381581b7022","Type":"ContainerStarted","Data":"6ae90da35a9a17509078df8ab980e3a84f21ba99008fc038307fe54f58ccd195"} Dec 06 03:46:17 crc kubenswrapper[4980]: I1206 03:46:17.709563 4980 generic.go:334] "Generic (PLEG): container finished" podID="e6d35c1e-b567-4b10-9c61-bc4d942f7a24" containerID="2df8c853db81a924ea6d68c7355b692bb22a355210a6f5cd0fa2157db87e35a9" exitCode=0 Dec 06 03:46:17 crc kubenswrapper[4980]: I1206 03:46:17.709637 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6hnct" event={"ID":"e6d35c1e-b567-4b10-9c61-bc4d942f7a24","Type":"ContainerDied","Data":"2df8c853db81a924ea6d68c7355b692bb22a355210a6f5cd0fa2157db87e35a9"} Dec 06 03:46:18 crc kubenswrapper[4980]: I1206 03:46:18.724486 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6hnct" event={"ID":"e6d35c1e-b567-4b10-9c61-bc4d942f7a24","Type":"ContainerStarted","Data":"146fff9923c5b2d3e6e57e6d52c202777f4618fc50ddcd2638bb474d25db0dfd"} Dec 06 03:46:18 crc kubenswrapper[4980]: I1206 03:46:18.724897 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6hnct" event={"ID":"e6d35c1e-b567-4b10-9c61-bc4d942f7a24","Type":"ContainerStarted","Data":"cb0547a9b518f667bfb32bcd0434e473a8c0a3dbfe74d6f2fe890e543653bdcd"} Dec 06 03:46:18 crc kubenswrapper[4980]: I1206 03:46:18.724919 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6hnct" event={"ID":"e6d35c1e-b567-4b10-9c61-bc4d942f7a24","Type":"ContainerStarted","Data":"ab124c6e1ca1ad8744e89d3e3659b4e7984f8d4c0a271ce29497390716f90f39"} Dec 06 03:46:18 crc kubenswrapper[4980]: I1206 03:46:18.724936 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6hnct" event={"ID":"e6d35c1e-b567-4b10-9c61-bc4d942f7a24","Type":"ContainerStarted","Data":"f4803bf169c69c361debe405a02478b66a9241da56903f2817bc4e88eb5d207d"} Dec 06 03:46:18 crc kubenswrapper[4980]: I1206 03:46:18.724957 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6hnct" event={"ID":"e6d35c1e-b567-4b10-9c61-bc4d942f7a24","Type":"ContainerStarted","Data":"dc8419bb9aa5a713b4aadd720b55eea67c3a4c20651f6b8c5c5e880ae88c3d9e"} Dec 06 03:46:18 crc kubenswrapper[4980]: I1206 03:46:18.724973 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6hnct" event={"ID":"e6d35c1e-b567-4b10-9c61-bc4d942f7a24","Type":"ContainerStarted","Data":"895e81eec4994846d042213e205dbba0b2ef232ad120a76bcb0a677b042acdb2"} Dec 06 03:46:21 crc kubenswrapper[4980]: I1206 03:46:21.760479 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6hnct" event={"ID":"e6d35c1e-b567-4b10-9c61-bc4d942f7a24","Type":"ContainerStarted","Data":"a97aad9de271645cd409c4750323ae0d42978275a45a7141858d7492812b74d0"} Dec 06 03:46:24 crc kubenswrapper[4980]: I1206 03:46:24.789151 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6hnct" event={"ID":"e6d35c1e-b567-4b10-9c61-bc4d942f7a24","Type":"ContainerStarted","Data":"0ccf5509c5e8351bd1ba343731634211ef9fabc7ae7bf8db91fb076cd04803e7"} Dec 06 03:46:24 crc kubenswrapper[4980]: I1206 03:46:24.789495 4980 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-6hnct" Dec 06 03:46:24 crc kubenswrapper[4980]: I1206 03:46:24.789708 4980 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-6hnct" Dec 06 03:46:24 crc kubenswrapper[4980]: I1206 03:46:24.789754 4980 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-6hnct" Dec 06 03:46:24 crc kubenswrapper[4980]: I1206 03:46:24.816611 4980 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-6hnct" Dec 06 03:46:24 crc kubenswrapper[4980]: I1206 03:46:24.818763 4980 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-6hnct" Dec 06 03:46:24 crc kubenswrapper[4980]: I1206 03:46:24.834124 4980 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-6hnct" podStartSLOduration=8.834067564 podStartE2EDuration="8.834067564s" podCreationTimestamp="2025-12-06 03:46:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 03:46:24.822280432 +0000 UTC m=+744.067031703" watchObservedRunningTime="2025-12-06 03:46:24.834067564 +0000 UTC m=+744.078818845" Dec 06 03:46:25 crc kubenswrapper[4980]: I1206 03:46:25.137319 4980 patch_prober.go:28] interesting pod/machine-config-daemon-r5zfb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 03:46:25 crc kubenswrapper[4980]: I1206 03:46:25.137446 4980 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-r5zfb" podUID="320f44d1-a671-4a91-b328-a8b0fdd8f23a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 03:46:41 crc kubenswrapper[4980]: I1206 03:46:41.920989 4980 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83brxgn"] Dec 06 03:46:41 crc kubenswrapper[4980]: I1206 03:46:41.924021 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83brxgn" Dec 06 03:46:41 crc kubenswrapper[4980]: I1206 03:46:41.928469 4980 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Dec 06 03:46:41 crc kubenswrapper[4980]: I1206 03:46:41.928756 4980 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83brxgn"] Dec 06 03:46:42 crc kubenswrapper[4980]: I1206 03:46:42.044288 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/73b1ef5a-140c-40df-a801-dc06cc56b5e2-bundle\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83brxgn\" (UID: \"73b1ef5a-140c-40df-a801-dc06cc56b5e2\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83brxgn" Dec 06 03:46:42 crc kubenswrapper[4980]: I1206 03:46:42.044665 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r5j8n\" (UniqueName: \"kubernetes.io/projected/73b1ef5a-140c-40df-a801-dc06cc56b5e2-kube-api-access-r5j8n\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83brxgn\" (UID: \"73b1ef5a-140c-40df-a801-dc06cc56b5e2\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83brxgn" Dec 06 03:46:42 crc kubenswrapper[4980]: I1206 03:46:42.044778 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/73b1ef5a-140c-40df-a801-dc06cc56b5e2-util\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83brxgn\" (UID: \"73b1ef5a-140c-40df-a801-dc06cc56b5e2\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83brxgn" Dec 06 03:46:42 crc kubenswrapper[4980]: I1206 03:46:42.145840 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r5j8n\" (UniqueName: \"kubernetes.io/projected/73b1ef5a-140c-40df-a801-dc06cc56b5e2-kube-api-access-r5j8n\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83brxgn\" (UID: \"73b1ef5a-140c-40df-a801-dc06cc56b5e2\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83brxgn" Dec 06 03:46:42 crc kubenswrapper[4980]: I1206 03:46:42.145891 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/73b1ef5a-140c-40df-a801-dc06cc56b5e2-util\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83brxgn\" (UID: \"73b1ef5a-140c-40df-a801-dc06cc56b5e2\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83brxgn" Dec 06 03:46:42 crc kubenswrapper[4980]: I1206 03:46:42.145916 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/73b1ef5a-140c-40df-a801-dc06cc56b5e2-bundle\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83brxgn\" (UID: \"73b1ef5a-140c-40df-a801-dc06cc56b5e2\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83brxgn" Dec 06 03:46:42 crc kubenswrapper[4980]: I1206 03:46:42.146431 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/73b1ef5a-140c-40df-a801-dc06cc56b5e2-bundle\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83brxgn\" (UID: \"73b1ef5a-140c-40df-a801-dc06cc56b5e2\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83brxgn" Dec 06 03:46:42 crc kubenswrapper[4980]: I1206 03:46:42.147279 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/73b1ef5a-140c-40df-a801-dc06cc56b5e2-util\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83brxgn\" (UID: \"73b1ef5a-140c-40df-a801-dc06cc56b5e2\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83brxgn" Dec 06 03:46:42 crc kubenswrapper[4980]: I1206 03:46:42.165605 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r5j8n\" (UniqueName: \"kubernetes.io/projected/73b1ef5a-140c-40df-a801-dc06cc56b5e2-kube-api-access-r5j8n\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83brxgn\" (UID: \"73b1ef5a-140c-40df-a801-dc06cc56b5e2\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83brxgn" Dec 06 03:46:42 crc kubenswrapper[4980]: I1206 03:46:42.288639 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83brxgn" Dec 06 03:46:42 crc kubenswrapper[4980]: I1206 03:46:42.500354 4980 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83brxgn"] Dec 06 03:46:42 crc kubenswrapper[4980]: W1206 03:46:42.510584 4980 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod73b1ef5a_140c_40df_a801_dc06cc56b5e2.slice/crio-40421a4597d90704352909f6c31bdb9d622f1f22fe9b62c62c8a159d284fdf42 WatchSource:0}: Error finding container 40421a4597d90704352909f6c31bdb9d622f1f22fe9b62c62c8a159d284fdf42: Status 404 returned error can't find the container with id 40421a4597d90704352909f6c31bdb9d622f1f22fe9b62c62c8a159d284fdf42 Dec 06 03:46:42 crc kubenswrapper[4980]: I1206 03:46:42.918818 4980 generic.go:334] "Generic (PLEG): container finished" podID="73b1ef5a-140c-40df-a801-dc06cc56b5e2" containerID="e73c80475d4826cfee1cab63deb29523048fd909dc16592f8f0f4b4dcdf9e09e" exitCode=0 Dec 06 03:46:42 crc kubenswrapper[4980]: I1206 03:46:42.918894 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83brxgn" event={"ID":"73b1ef5a-140c-40df-a801-dc06cc56b5e2","Type":"ContainerDied","Data":"e73c80475d4826cfee1cab63deb29523048fd909dc16592f8f0f4b4dcdf9e09e"} Dec 06 03:46:42 crc kubenswrapper[4980]: I1206 03:46:42.918935 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83brxgn" event={"ID":"73b1ef5a-140c-40df-a801-dc06cc56b5e2","Type":"ContainerStarted","Data":"40421a4597d90704352909f6c31bdb9d622f1f22fe9b62c62c8a159d284fdf42"} Dec 06 03:46:42 crc kubenswrapper[4980]: I1206 03:46:42.921888 4980 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 06 03:46:44 crc kubenswrapper[4980]: I1206 03:46:44.936340 4980 generic.go:334] "Generic (PLEG): container finished" podID="73b1ef5a-140c-40df-a801-dc06cc56b5e2" containerID="e4e1a4a72e00e3e1944fd3ee689d7d9ee60ca24250ef2cd22b18e384e521b4fd" exitCode=0 Dec 06 03:46:44 crc kubenswrapper[4980]: I1206 03:46:44.936494 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83brxgn" event={"ID":"73b1ef5a-140c-40df-a801-dc06cc56b5e2","Type":"ContainerDied","Data":"e4e1a4a72e00e3e1944fd3ee689d7d9ee60ca24250ef2cd22b18e384e521b4fd"} Dec 06 03:46:45 crc kubenswrapper[4980]: I1206 03:46:45.945289 4980 generic.go:334] "Generic (PLEG): container finished" podID="73b1ef5a-140c-40df-a801-dc06cc56b5e2" containerID="9f13d158b0fab89b6a8149d6e2bb3bd4924e56249500663ce10d5b9a4c83d600" exitCode=0 Dec 06 03:46:45 crc kubenswrapper[4980]: I1206 03:46:45.945354 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83brxgn" event={"ID":"73b1ef5a-140c-40df-a801-dc06cc56b5e2","Type":"ContainerDied","Data":"9f13d158b0fab89b6a8149d6e2bb3bd4924e56249500663ce10d5b9a4c83d600"} Dec 06 03:46:46 crc kubenswrapper[4980]: I1206 03:46:46.488341 4980 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-6hnct" Dec 06 03:46:47 crc kubenswrapper[4980]: I1206 03:46:47.256132 4980 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83brxgn" Dec 06 03:46:47 crc kubenswrapper[4980]: I1206 03:46:47.380656 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/73b1ef5a-140c-40df-a801-dc06cc56b5e2-bundle\") pod \"73b1ef5a-140c-40df-a801-dc06cc56b5e2\" (UID: \"73b1ef5a-140c-40df-a801-dc06cc56b5e2\") " Dec 06 03:46:47 crc kubenswrapper[4980]: I1206 03:46:47.380698 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/73b1ef5a-140c-40df-a801-dc06cc56b5e2-util\") pod \"73b1ef5a-140c-40df-a801-dc06cc56b5e2\" (UID: \"73b1ef5a-140c-40df-a801-dc06cc56b5e2\") " Dec 06 03:46:47 crc kubenswrapper[4980]: I1206 03:46:47.380797 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r5j8n\" (UniqueName: \"kubernetes.io/projected/73b1ef5a-140c-40df-a801-dc06cc56b5e2-kube-api-access-r5j8n\") pod \"73b1ef5a-140c-40df-a801-dc06cc56b5e2\" (UID: \"73b1ef5a-140c-40df-a801-dc06cc56b5e2\") " Dec 06 03:46:47 crc kubenswrapper[4980]: I1206 03:46:47.381756 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/73b1ef5a-140c-40df-a801-dc06cc56b5e2-bundle" (OuterVolumeSpecName: "bundle") pod "73b1ef5a-140c-40df-a801-dc06cc56b5e2" (UID: "73b1ef5a-140c-40df-a801-dc06cc56b5e2"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 03:46:47 crc kubenswrapper[4980]: I1206 03:46:47.386395 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/73b1ef5a-140c-40df-a801-dc06cc56b5e2-kube-api-access-r5j8n" (OuterVolumeSpecName: "kube-api-access-r5j8n") pod "73b1ef5a-140c-40df-a801-dc06cc56b5e2" (UID: "73b1ef5a-140c-40df-a801-dc06cc56b5e2"). InnerVolumeSpecName "kube-api-access-r5j8n". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 03:46:47 crc kubenswrapper[4980]: I1206 03:46:47.394279 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/73b1ef5a-140c-40df-a801-dc06cc56b5e2-util" (OuterVolumeSpecName: "util") pod "73b1ef5a-140c-40df-a801-dc06cc56b5e2" (UID: "73b1ef5a-140c-40df-a801-dc06cc56b5e2"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 03:46:47 crc kubenswrapper[4980]: I1206 03:46:47.481778 4980 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/73b1ef5a-140c-40df-a801-dc06cc56b5e2-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 03:46:47 crc kubenswrapper[4980]: I1206 03:46:47.481819 4980 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/73b1ef5a-140c-40df-a801-dc06cc56b5e2-util\") on node \"crc\" DevicePath \"\"" Dec 06 03:46:47 crc kubenswrapper[4980]: I1206 03:46:47.481832 4980 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r5j8n\" (UniqueName: \"kubernetes.io/projected/73b1ef5a-140c-40df-a801-dc06cc56b5e2-kube-api-access-r5j8n\") on node \"crc\" DevicePath \"\"" Dec 06 03:46:47 crc kubenswrapper[4980]: I1206 03:46:47.658116 4980 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-7bwr5"] Dec 06 03:46:47 crc kubenswrapper[4980]: E1206 03:46:47.658369 4980 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="73b1ef5a-140c-40df-a801-dc06cc56b5e2" containerName="pull" Dec 06 03:46:47 crc kubenswrapper[4980]: I1206 03:46:47.658395 4980 state_mem.go:107] "Deleted CPUSet assignment" podUID="73b1ef5a-140c-40df-a801-dc06cc56b5e2" containerName="pull" Dec 06 03:46:47 crc kubenswrapper[4980]: E1206 03:46:47.658418 4980 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="73b1ef5a-140c-40df-a801-dc06cc56b5e2" containerName="util" Dec 06 03:46:47 crc kubenswrapper[4980]: I1206 03:46:47.658427 4980 state_mem.go:107] "Deleted CPUSet assignment" podUID="73b1ef5a-140c-40df-a801-dc06cc56b5e2" containerName="util" Dec 06 03:46:47 crc kubenswrapper[4980]: E1206 03:46:47.658446 4980 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="73b1ef5a-140c-40df-a801-dc06cc56b5e2" containerName="extract" Dec 06 03:46:47 crc kubenswrapper[4980]: I1206 03:46:47.658456 4980 state_mem.go:107] "Deleted CPUSet assignment" podUID="73b1ef5a-140c-40df-a801-dc06cc56b5e2" containerName="extract" Dec 06 03:46:47 crc kubenswrapper[4980]: I1206 03:46:47.658635 4980 memory_manager.go:354] "RemoveStaleState removing state" podUID="73b1ef5a-140c-40df-a801-dc06cc56b5e2" containerName="extract" Dec 06 03:46:47 crc kubenswrapper[4980]: I1206 03:46:47.659676 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-7bwr5" Dec 06 03:46:47 crc kubenswrapper[4980]: I1206 03:46:47.668846 4980 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-7bwr5"] Dec 06 03:46:47 crc kubenswrapper[4980]: I1206 03:46:47.684647 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1a726fb7-975f-4b7b-9166-86785f01d360-utilities\") pod \"redhat-operators-7bwr5\" (UID: \"1a726fb7-975f-4b7b-9166-86785f01d360\") " pod="openshift-marketplace/redhat-operators-7bwr5" Dec 06 03:46:47 crc kubenswrapper[4980]: I1206 03:46:47.684710 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kkzf8\" (UniqueName: \"kubernetes.io/projected/1a726fb7-975f-4b7b-9166-86785f01d360-kube-api-access-kkzf8\") pod \"redhat-operators-7bwr5\" (UID: \"1a726fb7-975f-4b7b-9166-86785f01d360\") " pod="openshift-marketplace/redhat-operators-7bwr5" Dec 06 03:46:47 crc kubenswrapper[4980]: I1206 03:46:47.684754 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1a726fb7-975f-4b7b-9166-86785f01d360-catalog-content\") pod \"redhat-operators-7bwr5\" (UID: \"1a726fb7-975f-4b7b-9166-86785f01d360\") " pod="openshift-marketplace/redhat-operators-7bwr5" Dec 06 03:46:47 crc kubenswrapper[4980]: I1206 03:46:47.785898 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1a726fb7-975f-4b7b-9166-86785f01d360-utilities\") pod \"redhat-operators-7bwr5\" (UID: \"1a726fb7-975f-4b7b-9166-86785f01d360\") " pod="openshift-marketplace/redhat-operators-7bwr5" Dec 06 03:46:47 crc kubenswrapper[4980]: I1206 03:46:47.785951 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kkzf8\" (UniqueName: \"kubernetes.io/projected/1a726fb7-975f-4b7b-9166-86785f01d360-kube-api-access-kkzf8\") pod \"redhat-operators-7bwr5\" (UID: \"1a726fb7-975f-4b7b-9166-86785f01d360\") " pod="openshift-marketplace/redhat-operators-7bwr5" Dec 06 03:46:47 crc kubenswrapper[4980]: I1206 03:46:47.786000 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1a726fb7-975f-4b7b-9166-86785f01d360-catalog-content\") pod \"redhat-operators-7bwr5\" (UID: \"1a726fb7-975f-4b7b-9166-86785f01d360\") " pod="openshift-marketplace/redhat-operators-7bwr5" Dec 06 03:46:47 crc kubenswrapper[4980]: I1206 03:46:47.786644 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1a726fb7-975f-4b7b-9166-86785f01d360-utilities\") pod \"redhat-operators-7bwr5\" (UID: \"1a726fb7-975f-4b7b-9166-86785f01d360\") " pod="openshift-marketplace/redhat-operators-7bwr5" Dec 06 03:46:47 crc kubenswrapper[4980]: I1206 03:46:47.786748 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1a726fb7-975f-4b7b-9166-86785f01d360-catalog-content\") pod \"redhat-operators-7bwr5\" (UID: \"1a726fb7-975f-4b7b-9166-86785f01d360\") " pod="openshift-marketplace/redhat-operators-7bwr5" Dec 06 03:46:47 crc kubenswrapper[4980]: I1206 03:46:47.806324 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kkzf8\" (UniqueName: \"kubernetes.io/projected/1a726fb7-975f-4b7b-9166-86785f01d360-kube-api-access-kkzf8\") pod \"redhat-operators-7bwr5\" (UID: \"1a726fb7-975f-4b7b-9166-86785f01d360\") " pod="openshift-marketplace/redhat-operators-7bwr5" Dec 06 03:46:47 crc kubenswrapper[4980]: I1206 03:46:47.961807 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83brxgn" event={"ID":"73b1ef5a-140c-40df-a801-dc06cc56b5e2","Type":"ContainerDied","Data":"40421a4597d90704352909f6c31bdb9d622f1f22fe9b62c62c8a159d284fdf42"} Dec 06 03:46:47 crc kubenswrapper[4980]: I1206 03:46:47.961868 4980 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="40421a4597d90704352909f6c31bdb9d622f1f22fe9b62c62c8a159d284fdf42" Dec 06 03:46:47 crc kubenswrapper[4980]: I1206 03:46:47.961920 4980 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83brxgn" Dec 06 03:46:47 crc kubenswrapper[4980]: I1206 03:46:47.986117 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-7bwr5" Dec 06 03:46:48 crc kubenswrapper[4980]: I1206 03:46:48.209234 4980 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-7bwr5"] Dec 06 03:46:48 crc kubenswrapper[4980]: I1206 03:46:48.970707 4980 generic.go:334] "Generic (PLEG): container finished" podID="1a726fb7-975f-4b7b-9166-86785f01d360" containerID="60d155013919076d329bde7ed9114d8ff81aa834f680d9552e8c68f22e8a6d0f" exitCode=0 Dec 06 03:46:48 crc kubenswrapper[4980]: I1206 03:46:48.970829 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7bwr5" event={"ID":"1a726fb7-975f-4b7b-9166-86785f01d360","Type":"ContainerDied","Data":"60d155013919076d329bde7ed9114d8ff81aa834f680d9552e8c68f22e8a6d0f"} Dec 06 03:46:48 crc kubenswrapper[4980]: I1206 03:46:48.970903 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7bwr5" event={"ID":"1a726fb7-975f-4b7b-9166-86785f01d360","Type":"ContainerStarted","Data":"179ab0eb944163551b02a188110d1d58a134ca13bb7deb1c5370192e8690a822"} Dec 06 03:46:49 crc kubenswrapper[4980]: I1206 03:46:49.987214 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7bwr5" event={"ID":"1a726fb7-975f-4b7b-9166-86785f01d360","Type":"ContainerStarted","Data":"bbe74c02387b81c85c5f6cfe730b86ad6a1e5c149a59cdf09c8fab9b17428f1d"} Dec 06 03:46:50 crc kubenswrapper[4980]: I1206 03:46:50.996240 4980 generic.go:334] "Generic (PLEG): container finished" podID="1a726fb7-975f-4b7b-9166-86785f01d360" containerID="bbe74c02387b81c85c5f6cfe730b86ad6a1e5c149a59cdf09c8fab9b17428f1d" exitCode=0 Dec 06 03:46:50 crc kubenswrapper[4980]: I1206 03:46:50.996319 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7bwr5" event={"ID":"1a726fb7-975f-4b7b-9166-86785f01d360","Type":"ContainerDied","Data":"bbe74c02387b81c85c5f6cfe730b86ad6a1e5c149a59cdf09c8fab9b17428f1d"} Dec 06 03:46:52 crc kubenswrapper[4980]: I1206 03:46:52.005525 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7bwr5" event={"ID":"1a726fb7-975f-4b7b-9166-86785f01d360","Type":"ContainerStarted","Data":"d9d93975e1a2c27a75abbe3c0a2d6142a7d9de5296a487d15265accac97f7c84"} Dec 06 03:46:52 crc kubenswrapper[4980]: I1206 03:46:52.027124 4980 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-7bwr5" podStartSLOduration=2.563770556 podStartE2EDuration="5.027087574s" podCreationTimestamp="2025-12-06 03:46:47 +0000 UTC" firstStartedPulling="2025-12-06 03:46:48.978533575 +0000 UTC m=+768.223284846" lastFinishedPulling="2025-12-06 03:46:51.441850593 +0000 UTC m=+770.686601864" observedRunningTime="2025-12-06 03:46:52.023090728 +0000 UTC m=+771.267842009" watchObservedRunningTime="2025-12-06 03:46:52.027087574 +0000 UTC m=+771.271838845" Dec 06 03:46:52 crc kubenswrapper[4980]: I1206 03:46:52.415410 4980 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Dec 06 03:46:55 crc kubenswrapper[4980]: I1206 03:46:55.137708 4980 patch_prober.go:28] interesting pod/machine-config-daemon-r5zfb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 03:46:55 crc kubenswrapper[4980]: I1206 03:46:55.138068 4980 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-r5zfb" podUID="320f44d1-a671-4a91-b328-a8b0fdd8f23a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 03:46:57 crc kubenswrapper[4980]: I1206 03:46:57.191766 4980 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-controller-manager-555b967656-x878b"] Dec 06 03:46:57 crc kubenswrapper[4980]: I1206 03:46:57.192810 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-555b967656-x878b" Dec 06 03:46:57 crc kubenswrapper[4980]: I1206 03:46:57.196487 4980 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-controller-manager-service-cert" Dec 06 03:46:57 crc kubenswrapper[4980]: I1206 03:46:57.196772 4980 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"openshift-service-ca.crt" Dec 06 03:46:57 crc kubenswrapper[4980]: I1206 03:46:57.196968 4980 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-cert" Dec 06 03:46:57 crc kubenswrapper[4980]: I1206 03:46:57.197379 4980 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"manager-account-dockercfg-wbzgn" Dec 06 03:46:57 crc kubenswrapper[4980]: I1206 03:46:57.198526 4980 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"kube-root-ca.crt" Dec 06 03:46:57 crc kubenswrapper[4980]: I1206 03:46:57.205578 4980 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-555b967656-x878b"] Dec 06 03:46:57 crc kubenswrapper[4980]: I1206 03:46:57.360332 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/245c6ddf-2c16-419a-9931-f5887917d145-apiservice-cert\") pod \"metallb-operator-controller-manager-555b967656-x878b\" (UID: \"245c6ddf-2c16-419a-9931-f5887917d145\") " pod="metallb-system/metallb-operator-controller-manager-555b967656-x878b" Dec 06 03:46:57 crc kubenswrapper[4980]: I1206 03:46:57.360425 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6gxtq\" (UniqueName: \"kubernetes.io/projected/245c6ddf-2c16-419a-9931-f5887917d145-kube-api-access-6gxtq\") pod \"metallb-operator-controller-manager-555b967656-x878b\" (UID: \"245c6ddf-2c16-419a-9931-f5887917d145\") " pod="metallb-system/metallb-operator-controller-manager-555b967656-x878b" Dec 06 03:46:57 crc kubenswrapper[4980]: I1206 03:46:57.360492 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/245c6ddf-2c16-419a-9931-f5887917d145-webhook-cert\") pod \"metallb-operator-controller-manager-555b967656-x878b\" (UID: \"245c6ddf-2c16-419a-9931-f5887917d145\") " pod="metallb-system/metallb-operator-controller-manager-555b967656-x878b" Dec 06 03:46:57 crc kubenswrapper[4980]: I1206 03:46:57.447377 4980 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-webhook-server-86db4b77c-4tjd7"] Dec 06 03:46:57 crc kubenswrapper[4980]: I1206 03:46:57.448170 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-86db4b77c-4tjd7" Dec 06 03:46:57 crc kubenswrapper[4980]: I1206 03:46:57.453375 4980 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-dockercfg-lw46q" Dec 06 03:46:57 crc kubenswrapper[4980]: I1206 03:46:57.453528 4980 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-service-cert" Dec 06 03:46:57 crc kubenswrapper[4980]: I1206 03:46:57.454324 4980 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Dec 06 03:46:57 crc kubenswrapper[4980]: I1206 03:46:57.461682 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/245c6ddf-2c16-419a-9931-f5887917d145-webhook-cert\") pod \"metallb-operator-controller-manager-555b967656-x878b\" (UID: \"245c6ddf-2c16-419a-9931-f5887917d145\") " pod="metallb-system/metallb-operator-controller-manager-555b967656-x878b" Dec 06 03:46:57 crc kubenswrapper[4980]: I1206 03:46:57.461741 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/245c6ddf-2c16-419a-9931-f5887917d145-apiservice-cert\") pod \"metallb-operator-controller-manager-555b967656-x878b\" (UID: \"245c6ddf-2c16-419a-9931-f5887917d145\") " pod="metallb-system/metallb-operator-controller-manager-555b967656-x878b" Dec 06 03:46:57 crc kubenswrapper[4980]: I1206 03:46:57.461793 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6gxtq\" (UniqueName: \"kubernetes.io/projected/245c6ddf-2c16-419a-9931-f5887917d145-kube-api-access-6gxtq\") pod \"metallb-operator-controller-manager-555b967656-x878b\" (UID: \"245c6ddf-2c16-419a-9931-f5887917d145\") " pod="metallb-system/metallb-operator-controller-manager-555b967656-x878b" Dec 06 03:46:57 crc kubenswrapper[4980]: I1206 03:46:57.474913 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/245c6ddf-2c16-419a-9931-f5887917d145-webhook-cert\") pod \"metallb-operator-controller-manager-555b967656-x878b\" (UID: \"245c6ddf-2c16-419a-9931-f5887917d145\") " pod="metallb-system/metallb-operator-controller-manager-555b967656-x878b" Dec 06 03:46:57 crc kubenswrapper[4980]: I1206 03:46:57.476690 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/245c6ddf-2c16-419a-9931-f5887917d145-apiservice-cert\") pod \"metallb-operator-controller-manager-555b967656-x878b\" (UID: \"245c6ddf-2c16-419a-9931-f5887917d145\") " pod="metallb-system/metallb-operator-controller-manager-555b967656-x878b" Dec 06 03:46:57 crc kubenswrapper[4980]: I1206 03:46:57.482164 4980 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-86db4b77c-4tjd7"] Dec 06 03:46:57 crc kubenswrapper[4980]: I1206 03:46:57.489698 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6gxtq\" (UniqueName: \"kubernetes.io/projected/245c6ddf-2c16-419a-9931-f5887917d145-kube-api-access-6gxtq\") pod \"metallb-operator-controller-manager-555b967656-x878b\" (UID: \"245c6ddf-2c16-419a-9931-f5887917d145\") " pod="metallb-system/metallb-operator-controller-manager-555b967656-x878b" Dec 06 03:46:57 crc kubenswrapper[4980]: I1206 03:46:57.563339 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xhlvj\" (UniqueName: \"kubernetes.io/projected/7e314ba1-bb63-4da4-b43c-97f9bc74b99f-kube-api-access-xhlvj\") pod \"metallb-operator-webhook-server-86db4b77c-4tjd7\" (UID: \"7e314ba1-bb63-4da4-b43c-97f9bc74b99f\") " pod="metallb-system/metallb-operator-webhook-server-86db4b77c-4tjd7" Dec 06 03:46:57 crc kubenswrapper[4980]: I1206 03:46:57.563416 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/7e314ba1-bb63-4da4-b43c-97f9bc74b99f-apiservice-cert\") pod \"metallb-operator-webhook-server-86db4b77c-4tjd7\" (UID: \"7e314ba1-bb63-4da4-b43c-97f9bc74b99f\") " pod="metallb-system/metallb-operator-webhook-server-86db4b77c-4tjd7" Dec 06 03:46:57 crc kubenswrapper[4980]: I1206 03:46:57.563439 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/7e314ba1-bb63-4da4-b43c-97f9bc74b99f-webhook-cert\") pod \"metallb-operator-webhook-server-86db4b77c-4tjd7\" (UID: \"7e314ba1-bb63-4da4-b43c-97f9bc74b99f\") " pod="metallb-system/metallb-operator-webhook-server-86db4b77c-4tjd7" Dec 06 03:46:57 crc kubenswrapper[4980]: I1206 03:46:57.566242 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-555b967656-x878b" Dec 06 03:46:57 crc kubenswrapper[4980]: I1206 03:46:57.664219 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/7e314ba1-bb63-4da4-b43c-97f9bc74b99f-webhook-cert\") pod \"metallb-operator-webhook-server-86db4b77c-4tjd7\" (UID: \"7e314ba1-bb63-4da4-b43c-97f9bc74b99f\") " pod="metallb-system/metallb-operator-webhook-server-86db4b77c-4tjd7" Dec 06 03:46:57 crc kubenswrapper[4980]: I1206 03:46:57.664352 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xhlvj\" (UniqueName: \"kubernetes.io/projected/7e314ba1-bb63-4da4-b43c-97f9bc74b99f-kube-api-access-xhlvj\") pod \"metallb-operator-webhook-server-86db4b77c-4tjd7\" (UID: \"7e314ba1-bb63-4da4-b43c-97f9bc74b99f\") " pod="metallb-system/metallb-operator-webhook-server-86db4b77c-4tjd7" Dec 06 03:46:57 crc kubenswrapper[4980]: I1206 03:46:57.664428 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/7e314ba1-bb63-4da4-b43c-97f9bc74b99f-apiservice-cert\") pod \"metallb-operator-webhook-server-86db4b77c-4tjd7\" (UID: \"7e314ba1-bb63-4da4-b43c-97f9bc74b99f\") " pod="metallb-system/metallb-operator-webhook-server-86db4b77c-4tjd7" Dec 06 03:46:57 crc kubenswrapper[4980]: I1206 03:46:57.668431 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/7e314ba1-bb63-4da4-b43c-97f9bc74b99f-apiservice-cert\") pod \"metallb-operator-webhook-server-86db4b77c-4tjd7\" (UID: \"7e314ba1-bb63-4da4-b43c-97f9bc74b99f\") " pod="metallb-system/metallb-operator-webhook-server-86db4b77c-4tjd7" Dec 06 03:46:57 crc kubenswrapper[4980]: I1206 03:46:57.671207 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/7e314ba1-bb63-4da4-b43c-97f9bc74b99f-webhook-cert\") pod \"metallb-operator-webhook-server-86db4b77c-4tjd7\" (UID: \"7e314ba1-bb63-4da4-b43c-97f9bc74b99f\") " pod="metallb-system/metallb-operator-webhook-server-86db4b77c-4tjd7" Dec 06 03:46:57 crc kubenswrapper[4980]: I1206 03:46:57.688014 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xhlvj\" (UniqueName: \"kubernetes.io/projected/7e314ba1-bb63-4da4-b43c-97f9bc74b99f-kube-api-access-xhlvj\") pod \"metallb-operator-webhook-server-86db4b77c-4tjd7\" (UID: \"7e314ba1-bb63-4da4-b43c-97f9bc74b99f\") " pod="metallb-system/metallb-operator-webhook-server-86db4b77c-4tjd7" Dec 06 03:46:57 crc kubenswrapper[4980]: I1206 03:46:57.766173 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-86db4b77c-4tjd7" Dec 06 03:46:57 crc kubenswrapper[4980]: I1206 03:46:57.880688 4980 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-555b967656-x878b"] Dec 06 03:46:57 crc kubenswrapper[4980]: I1206 03:46:57.987149 4980 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-7bwr5" Dec 06 03:46:57 crc kubenswrapper[4980]: I1206 03:46:57.987207 4980 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-7bwr5" Dec 06 03:46:58 crc kubenswrapper[4980]: I1206 03:46:58.049423 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-555b967656-x878b" event={"ID":"245c6ddf-2c16-419a-9931-f5887917d145","Type":"ContainerStarted","Data":"168d3402f6e2a9cbb6317641df02ec5a61f728147333b18e145611131a6d8e46"} Dec 06 03:46:58 crc kubenswrapper[4980]: I1206 03:46:58.058330 4980 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-7bwr5" Dec 06 03:46:58 crc kubenswrapper[4980]: I1206 03:46:58.085147 4980 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-86db4b77c-4tjd7"] Dec 06 03:46:58 crc kubenswrapper[4980]: W1206 03:46:58.091964 4980 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7e314ba1_bb63_4da4_b43c_97f9bc74b99f.slice/crio-13712bc20167a45de206e16667e481a77c77d4e04193a2533719f6f250fcd792 WatchSource:0}: Error finding container 13712bc20167a45de206e16667e481a77c77d4e04193a2533719f6f250fcd792: Status 404 returned error can't find the container with id 13712bc20167a45de206e16667e481a77c77d4e04193a2533719f6f250fcd792 Dec 06 03:46:58 crc kubenswrapper[4980]: I1206 03:46:58.108812 4980 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-7bwr5" Dec 06 03:46:59 crc kubenswrapper[4980]: I1206 03:46:59.056151 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-86db4b77c-4tjd7" event={"ID":"7e314ba1-bb63-4da4-b43c-97f9bc74b99f","Type":"ContainerStarted","Data":"13712bc20167a45de206e16667e481a77c77d4e04193a2533719f6f250fcd792"} Dec 06 03:47:00 crc kubenswrapper[4980]: I1206 03:47:00.468723 4980 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-7bwr5"] Dec 06 03:47:00 crc kubenswrapper[4980]: I1206 03:47:00.469325 4980 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-7bwr5" podUID="1a726fb7-975f-4b7b-9166-86785f01d360" containerName="registry-server" containerID="cri-o://d9d93975e1a2c27a75abbe3c0a2d6142a7d9de5296a487d15265accac97f7c84" gracePeriod=2 Dec 06 03:47:02 crc kubenswrapper[4980]: I1206 03:47:02.085212 4980 generic.go:334] "Generic (PLEG): container finished" podID="1a726fb7-975f-4b7b-9166-86785f01d360" containerID="d9d93975e1a2c27a75abbe3c0a2d6142a7d9de5296a487d15265accac97f7c84" exitCode=0 Dec 06 03:47:02 crc kubenswrapper[4980]: I1206 03:47:02.085308 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7bwr5" event={"ID":"1a726fb7-975f-4b7b-9166-86785f01d360","Type":"ContainerDied","Data":"d9d93975e1a2c27a75abbe3c0a2d6142a7d9de5296a487d15265accac97f7c84"} Dec 06 03:47:04 crc kubenswrapper[4980]: I1206 03:47:04.093045 4980 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-7bwr5" Dec 06 03:47:04 crc kubenswrapper[4980]: I1206 03:47:04.106501 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7bwr5" event={"ID":"1a726fb7-975f-4b7b-9166-86785f01d360","Type":"ContainerDied","Data":"179ab0eb944163551b02a188110d1d58a134ca13bb7deb1c5370192e8690a822"} Dec 06 03:47:04 crc kubenswrapper[4980]: I1206 03:47:04.106601 4980 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-7bwr5" Dec 06 03:47:04 crc kubenswrapper[4980]: I1206 03:47:04.106684 4980 scope.go:117] "RemoveContainer" containerID="d9d93975e1a2c27a75abbe3c0a2d6142a7d9de5296a487d15265accac97f7c84" Dec 06 03:47:04 crc kubenswrapper[4980]: I1206 03:47:04.238618 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1a726fb7-975f-4b7b-9166-86785f01d360-catalog-content\") pod \"1a726fb7-975f-4b7b-9166-86785f01d360\" (UID: \"1a726fb7-975f-4b7b-9166-86785f01d360\") " Dec 06 03:47:04 crc kubenswrapper[4980]: I1206 03:47:04.238731 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1a726fb7-975f-4b7b-9166-86785f01d360-utilities\") pod \"1a726fb7-975f-4b7b-9166-86785f01d360\" (UID: \"1a726fb7-975f-4b7b-9166-86785f01d360\") " Dec 06 03:47:04 crc kubenswrapper[4980]: I1206 03:47:04.238788 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kkzf8\" (UniqueName: \"kubernetes.io/projected/1a726fb7-975f-4b7b-9166-86785f01d360-kube-api-access-kkzf8\") pod \"1a726fb7-975f-4b7b-9166-86785f01d360\" (UID: \"1a726fb7-975f-4b7b-9166-86785f01d360\") " Dec 06 03:47:04 crc kubenswrapper[4980]: I1206 03:47:04.239816 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1a726fb7-975f-4b7b-9166-86785f01d360-utilities" (OuterVolumeSpecName: "utilities") pod "1a726fb7-975f-4b7b-9166-86785f01d360" (UID: "1a726fb7-975f-4b7b-9166-86785f01d360"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 03:47:04 crc kubenswrapper[4980]: I1206 03:47:04.243735 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1a726fb7-975f-4b7b-9166-86785f01d360-kube-api-access-kkzf8" (OuterVolumeSpecName: "kube-api-access-kkzf8") pod "1a726fb7-975f-4b7b-9166-86785f01d360" (UID: "1a726fb7-975f-4b7b-9166-86785f01d360"). InnerVolumeSpecName "kube-api-access-kkzf8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 03:47:04 crc kubenswrapper[4980]: I1206 03:47:04.340209 4980 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kkzf8\" (UniqueName: \"kubernetes.io/projected/1a726fb7-975f-4b7b-9166-86785f01d360-kube-api-access-kkzf8\") on node \"crc\" DevicePath \"\"" Dec 06 03:47:04 crc kubenswrapper[4980]: I1206 03:47:04.340249 4980 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1a726fb7-975f-4b7b-9166-86785f01d360-utilities\") on node \"crc\" DevicePath \"\"" Dec 06 03:47:04 crc kubenswrapper[4980]: I1206 03:47:04.367834 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1a726fb7-975f-4b7b-9166-86785f01d360-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1a726fb7-975f-4b7b-9166-86785f01d360" (UID: "1a726fb7-975f-4b7b-9166-86785f01d360"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 03:47:04 crc kubenswrapper[4980]: I1206 03:47:04.440982 4980 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1a726fb7-975f-4b7b-9166-86785f01d360-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 06 03:47:04 crc kubenswrapper[4980]: I1206 03:47:04.443296 4980 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-7bwr5"] Dec 06 03:47:04 crc kubenswrapper[4980]: I1206 03:47:04.447641 4980 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-7bwr5"] Dec 06 03:47:05 crc kubenswrapper[4980]: I1206 03:47:05.191778 4980 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1a726fb7-975f-4b7b-9166-86785f01d360" path="/var/lib/kubelet/pods/1a726fb7-975f-4b7b-9166-86785f01d360/volumes" Dec 06 03:47:05 crc kubenswrapper[4980]: I1206 03:47:05.540504 4980 scope.go:117] "RemoveContainer" containerID="bbe74c02387b81c85c5f6cfe730b86ad6a1e5c149a59cdf09c8fab9b17428f1d" Dec 06 03:47:05 crc kubenswrapper[4980]: I1206 03:47:05.573388 4980 scope.go:117] "RemoveContainer" containerID="60d155013919076d329bde7ed9114d8ff81aa834f680d9552e8c68f22e8a6d0f" Dec 06 03:47:06 crc kubenswrapper[4980]: I1206 03:47:06.146079 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-555b967656-x878b" event={"ID":"245c6ddf-2c16-419a-9931-f5887917d145","Type":"ContainerStarted","Data":"24a144e182ff2e0f366a82ad9b6795001655029ad2283965827660e520cb2ab1"} Dec 06 03:47:06 crc kubenswrapper[4980]: I1206 03:47:06.146742 4980 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-controller-manager-555b967656-x878b" Dec 06 03:47:06 crc kubenswrapper[4980]: I1206 03:47:06.147602 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-86db4b77c-4tjd7" event={"ID":"7e314ba1-bb63-4da4-b43c-97f9bc74b99f","Type":"ContainerStarted","Data":"301ca6080a5b1b487ec43a5f035060b6401b3cfcb98a4f1adcdf4eb34fb8caa5"} Dec 06 03:47:06 crc kubenswrapper[4980]: I1206 03:47:06.176221 4980 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-controller-manager-555b967656-x878b" podStartSLOduration=1.516549227 podStartE2EDuration="9.176179584s" podCreationTimestamp="2025-12-06 03:46:57 +0000 UTC" firstStartedPulling="2025-12-06 03:46:57.891068747 +0000 UTC m=+777.135820018" lastFinishedPulling="2025-12-06 03:47:05.550699104 +0000 UTC m=+784.795450375" observedRunningTime="2025-12-06 03:47:06.171454317 +0000 UTC m=+785.416205608" watchObservedRunningTime="2025-12-06 03:47:06.176179584 +0000 UTC m=+785.420930845" Dec 06 03:47:06 crc kubenswrapper[4980]: I1206 03:47:06.190231 4980 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-webhook-server-86db4b77c-4tjd7" podStartSLOduration=1.71163963 podStartE2EDuration="9.190211259s" podCreationTimestamp="2025-12-06 03:46:57 +0000 UTC" firstStartedPulling="2025-12-06 03:46:58.096482079 +0000 UTC m=+777.341233350" lastFinishedPulling="2025-12-06 03:47:05.575053708 +0000 UTC m=+784.819804979" observedRunningTime="2025-12-06 03:47:06.188683605 +0000 UTC m=+785.433434886" watchObservedRunningTime="2025-12-06 03:47:06.190211259 +0000 UTC m=+785.434962530" Dec 06 03:47:07 crc kubenswrapper[4980]: I1206 03:47:07.155053 4980 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-webhook-server-86db4b77c-4tjd7" Dec 06 03:47:17 crc kubenswrapper[4980]: I1206 03:47:17.780524 4980 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-webhook-server-86db4b77c-4tjd7" Dec 06 03:47:25 crc kubenswrapper[4980]: I1206 03:47:25.138091 4980 patch_prober.go:28] interesting pod/machine-config-daemon-r5zfb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 03:47:25 crc kubenswrapper[4980]: I1206 03:47:25.139061 4980 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-r5zfb" podUID="320f44d1-a671-4a91-b328-a8b0fdd8f23a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 03:47:25 crc kubenswrapper[4980]: I1206 03:47:25.139160 4980 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-r5zfb" Dec 06 03:47:25 crc kubenswrapper[4980]: I1206 03:47:25.140153 4980 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"c1b55d544de848b8b12ac6ab1d4d7ce58e20980d22589239fd7500d5b466d62e"} pod="openshift-machine-config-operator/machine-config-daemon-r5zfb" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 06 03:47:25 crc kubenswrapper[4980]: I1206 03:47:25.140259 4980 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-r5zfb" podUID="320f44d1-a671-4a91-b328-a8b0fdd8f23a" containerName="machine-config-daemon" containerID="cri-o://c1b55d544de848b8b12ac6ab1d4d7ce58e20980d22589239fd7500d5b466d62e" gracePeriod=600 Dec 06 03:47:26 crc kubenswrapper[4980]: I1206 03:47:26.290281 4980 generic.go:334] "Generic (PLEG): container finished" podID="320f44d1-a671-4a91-b328-a8b0fdd8f23a" containerID="c1b55d544de848b8b12ac6ab1d4d7ce58e20980d22589239fd7500d5b466d62e" exitCode=0 Dec 06 03:47:26 crc kubenswrapper[4980]: I1206 03:47:26.290393 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-r5zfb" event={"ID":"320f44d1-a671-4a91-b328-a8b0fdd8f23a","Type":"ContainerDied","Data":"c1b55d544de848b8b12ac6ab1d4d7ce58e20980d22589239fd7500d5b466d62e"} Dec 06 03:47:26 crc kubenswrapper[4980]: I1206 03:47:26.290852 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-r5zfb" event={"ID":"320f44d1-a671-4a91-b328-a8b0fdd8f23a","Type":"ContainerStarted","Data":"89e5bc79489f00d58aeba62f0124b8200c624808c64c4cf50ac8bd11cba63d00"} Dec 06 03:47:26 crc kubenswrapper[4980]: I1206 03:47:26.290892 4980 scope.go:117] "RemoveContainer" containerID="ca8484c03e9f51ca249527d107dc686e5fd267b68f8d80ac4706ca616b6cf584" Dec 06 03:47:37 crc kubenswrapper[4980]: I1206 03:47:37.571152 4980 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-controller-manager-555b967656-x878b" Dec 06 03:47:38 crc kubenswrapper[4980]: I1206 03:47:38.279450 4980 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-jh6gj"] Dec 06 03:47:38 crc kubenswrapper[4980]: E1206 03:47:38.280058 4980 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1a726fb7-975f-4b7b-9166-86785f01d360" containerName="extract-utilities" Dec 06 03:47:38 crc kubenswrapper[4980]: I1206 03:47:38.280082 4980 state_mem.go:107] "Deleted CPUSet assignment" podUID="1a726fb7-975f-4b7b-9166-86785f01d360" containerName="extract-utilities" Dec 06 03:47:38 crc kubenswrapper[4980]: E1206 03:47:38.280097 4980 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1a726fb7-975f-4b7b-9166-86785f01d360" containerName="extract-content" Dec 06 03:47:38 crc kubenswrapper[4980]: I1206 03:47:38.280107 4980 state_mem.go:107] "Deleted CPUSet assignment" podUID="1a726fb7-975f-4b7b-9166-86785f01d360" containerName="extract-content" Dec 06 03:47:38 crc kubenswrapper[4980]: E1206 03:47:38.280124 4980 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1a726fb7-975f-4b7b-9166-86785f01d360" containerName="registry-server" Dec 06 03:47:38 crc kubenswrapper[4980]: I1206 03:47:38.280132 4980 state_mem.go:107] "Deleted CPUSet assignment" podUID="1a726fb7-975f-4b7b-9166-86785f01d360" containerName="registry-server" Dec 06 03:47:38 crc kubenswrapper[4980]: I1206 03:47:38.280296 4980 memory_manager.go:354] "RemoveStaleState removing state" podUID="1a726fb7-975f-4b7b-9166-86785f01d360" containerName="registry-server" Dec 06 03:47:38 crc kubenswrapper[4980]: I1206 03:47:38.282631 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-jh6gj" Dec 06 03:47:38 crc kubenswrapper[4980]: I1206 03:47:38.284393 4980 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-certs-secret" Dec 06 03:47:38 crc kubenswrapper[4980]: I1206 03:47:38.284914 4980 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"frr-startup" Dec 06 03:47:38 crc kubenswrapper[4980]: I1206 03:47:38.285086 4980 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-daemon-dockercfg-9qvw9" Dec 06 03:47:38 crc kubenswrapper[4980]: I1206 03:47:38.287126 4980 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-webhook-server-7fcb986d4-7qtqr"] Dec 06 03:47:38 crc kubenswrapper[4980]: I1206 03:47:38.288067 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-7qtqr" Dec 06 03:47:38 crc kubenswrapper[4980]: I1206 03:47:38.290047 4980 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-webhook-server-cert" Dec 06 03:47:38 crc kubenswrapper[4980]: I1206 03:47:38.306339 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/97b725b0-8d1d-4c85-8c96-f96e269da5c5-reloader\") pod \"frr-k8s-jh6gj\" (UID: \"97b725b0-8d1d-4c85-8c96-f96e269da5c5\") " pod="metallb-system/frr-k8s-jh6gj" Dec 06 03:47:38 crc kubenswrapper[4980]: I1206 03:47:38.306416 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fx56t\" (UniqueName: \"kubernetes.io/projected/97b725b0-8d1d-4c85-8c96-f96e269da5c5-kube-api-access-fx56t\") pod \"frr-k8s-jh6gj\" (UID: \"97b725b0-8d1d-4c85-8c96-f96e269da5c5\") " pod="metallb-system/frr-k8s-jh6gj" Dec 06 03:47:38 crc kubenswrapper[4980]: I1206 03:47:38.306447 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/97b725b0-8d1d-4c85-8c96-f96e269da5c5-metrics\") pod \"frr-k8s-jh6gj\" (UID: \"97b725b0-8d1d-4c85-8c96-f96e269da5c5\") " pod="metallb-system/frr-k8s-jh6gj" Dec 06 03:47:38 crc kubenswrapper[4980]: I1206 03:47:38.306476 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/97b725b0-8d1d-4c85-8c96-f96e269da5c5-frr-conf\") pod \"frr-k8s-jh6gj\" (UID: \"97b725b0-8d1d-4c85-8c96-f96e269da5c5\") " pod="metallb-system/frr-k8s-jh6gj" Dec 06 03:47:38 crc kubenswrapper[4980]: I1206 03:47:38.306560 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/97b725b0-8d1d-4c85-8c96-f96e269da5c5-frr-sockets\") pod \"frr-k8s-jh6gj\" (UID: \"97b725b0-8d1d-4c85-8c96-f96e269da5c5\") " pod="metallb-system/frr-k8s-jh6gj" Dec 06 03:47:38 crc kubenswrapper[4980]: I1206 03:47:38.306584 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/97b725b0-8d1d-4c85-8c96-f96e269da5c5-frr-startup\") pod \"frr-k8s-jh6gj\" (UID: \"97b725b0-8d1d-4c85-8c96-f96e269da5c5\") " pod="metallb-system/frr-k8s-jh6gj" Dec 06 03:47:38 crc kubenswrapper[4980]: I1206 03:47:38.306608 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/1f860039-c5ba-4ba9-90b1-ecb00110015f-cert\") pod \"frr-k8s-webhook-server-7fcb986d4-7qtqr\" (UID: \"1f860039-c5ba-4ba9-90b1-ecb00110015f\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-7qtqr" Dec 06 03:47:38 crc kubenswrapper[4980]: I1206 03:47:38.306680 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z6g9k\" (UniqueName: \"kubernetes.io/projected/1f860039-c5ba-4ba9-90b1-ecb00110015f-kube-api-access-z6g9k\") pod \"frr-k8s-webhook-server-7fcb986d4-7qtqr\" (UID: \"1f860039-c5ba-4ba9-90b1-ecb00110015f\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-7qtqr" Dec 06 03:47:38 crc kubenswrapper[4980]: I1206 03:47:38.306704 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/97b725b0-8d1d-4c85-8c96-f96e269da5c5-metrics-certs\") pod \"frr-k8s-jh6gj\" (UID: \"97b725b0-8d1d-4c85-8c96-f96e269da5c5\") " pod="metallb-system/frr-k8s-jh6gj" Dec 06 03:47:38 crc kubenswrapper[4980]: I1206 03:47:38.312327 4980 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-7fcb986d4-7qtqr"] Dec 06 03:47:38 crc kubenswrapper[4980]: I1206 03:47:38.370924 4980 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/speaker-6wv7f"] Dec 06 03:47:38 crc kubenswrapper[4980]: I1206 03:47:38.372009 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-6wv7f" Dec 06 03:47:38 crc kubenswrapper[4980]: I1206 03:47:38.409209 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/97b725b0-8d1d-4c85-8c96-f96e269da5c5-frr-sockets\") pod \"frr-k8s-jh6gj\" (UID: \"97b725b0-8d1d-4c85-8c96-f96e269da5c5\") " pod="metallb-system/frr-k8s-jh6gj" Dec 06 03:47:38 crc kubenswrapper[4980]: I1206 03:47:38.409248 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/97b725b0-8d1d-4c85-8c96-f96e269da5c5-frr-startup\") pod \"frr-k8s-jh6gj\" (UID: \"97b725b0-8d1d-4c85-8c96-f96e269da5c5\") " pod="metallb-system/frr-k8s-jh6gj" Dec 06 03:47:38 crc kubenswrapper[4980]: I1206 03:47:38.409265 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/1f860039-c5ba-4ba9-90b1-ecb00110015f-cert\") pod \"frr-k8s-webhook-server-7fcb986d4-7qtqr\" (UID: \"1f860039-c5ba-4ba9-90b1-ecb00110015f\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-7qtqr" Dec 06 03:47:38 crc kubenswrapper[4980]: I1206 03:47:38.410298 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/97b725b0-8d1d-4c85-8c96-f96e269da5c5-frr-startup\") pod \"frr-k8s-jh6gj\" (UID: \"97b725b0-8d1d-4c85-8c96-f96e269da5c5\") " pod="metallb-system/frr-k8s-jh6gj" Dec 06 03:47:38 crc kubenswrapper[4980]: E1206 03:47:38.410376 4980 secret.go:188] Couldn't get secret metallb-system/frr-k8s-webhook-server-cert: secret "frr-k8s-webhook-server-cert" not found Dec 06 03:47:38 crc kubenswrapper[4980]: E1206 03:47:38.410432 4980 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/1f860039-c5ba-4ba9-90b1-ecb00110015f-cert podName:1f860039-c5ba-4ba9-90b1-ecb00110015f nodeName:}" failed. No retries permitted until 2025-12-06 03:47:38.910418741 +0000 UTC m=+818.155170002 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/1f860039-c5ba-4ba9-90b1-ecb00110015f-cert") pod "frr-k8s-webhook-server-7fcb986d4-7qtqr" (UID: "1f860039-c5ba-4ba9-90b1-ecb00110015f") : secret "frr-k8s-webhook-server-cert" not found Dec 06 03:47:38 crc kubenswrapper[4980]: I1206 03:47:38.411489 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/97b725b0-8d1d-4c85-8c96-f96e269da5c5-metrics-certs\") pod \"frr-k8s-jh6gj\" (UID: \"97b725b0-8d1d-4c85-8c96-f96e269da5c5\") " pod="metallb-system/frr-k8s-jh6gj" Dec 06 03:47:38 crc kubenswrapper[4980]: I1206 03:47:38.411533 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z6g9k\" (UniqueName: \"kubernetes.io/projected/1f860039-c5ba-4ba9-90b1-ecb00110015f-kube-api-access-z6g9k\") pod \"frr-k8s-webhook-server-7fcb986d4-7qtqr\" (UID: \"1f860039-c5ba-4ba9-90b1-ecb00110015f\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-7qtqr" Dec 06 03:47:38 crc kubenswrapper[4980]: I1206 03:47:38.411599 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/97b725b0-8d1d-4c85-8c96-f96e269da5c5-reloader\") pod \"frr-k8s-jh6gj\" (UID: \"97b725b0-8d1d-4c85-8c96-f96e269da5c5\") " pod="metallb-system/frr-k8s-jh6gj" Dec 06 03:47:38 crc kubenswrapper[4980]: I1206 03:47:38.411634 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fx56t\" (UniqueName: \"kubernetes.io/projected/97b725b0-8d1d-4c85-8c96-f96e269da5c5-kube-api-access-fx56t\") pod \"frr-k8s-jh6gj\" (UID: \"97b725b0-8d1d-4c85-8c96-f96e269da5c5\") " pod="metallb-system/frr-k8s-jh6gj" Dec 06 03:47:38 crc kubenswrapper[4980]: I1206 03:47:38.411654 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/97b725b0-8d1d-4c85-8c96-f96e269da5c5-metrics\") pod \"frr-k8s-jh6gj\" (UID: \"97b725b0-8d1d-4c85-8c96-f96e269da5c5\") " pod="metallb-system/frr-k8s-jh6gj" Dec 06 03:47:38 crc kubenswrapper[4980]: I1206 03:47:38.411673 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/97b725b0-8d1d-4c85-8c96-f96e269da5c5-frr-conf\") pod \"frr-k8s-jh6gj\" (UID: \"97b725b0-8d1d-4c85-8c96-f96e269da5c5\") " pod="metallb-system/frr-k8s-jh6gj" Dec 06 03:47:38 crc kubenswrapper[4980]: I1206 03:47:38.412560 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/97b725b0-8d1d-4c85-8c96-f96e269da5c5-frr-conf\") pod \"frr-k8s-jh6gj\" (UID: \"97b725b0-8d1d-4c85-8c96-f96e269da5c5\") " pod="metallb-system/frr-k8s-jh6gj" Dec 06 03:47:38 crc kubenswrapper[4980]: I1206 03:47:38.412874 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/97b725b0-8d1d-4c85-8c96-f96e269da5c5-reloader\") pod \"frr-k8s-jh6gj\" (UID: \"97b725b0-8d1d-4c85-8c96-f96e269da5c5\") " pod="metallb-system/frr-k8s-jh6gj" Dec 06 03:47:38 crc kubenswrapper[4980]: I1206 03:47:38.413031 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/97b725b0-8d1d-4c85-8c96-f96e269da5c5-metrics\") pod \"frr-k8s-jh6gj\" (UID: \"97b725b0-8d1d-4c85-8c96-f96e269da5c5\") " pod="metallb-system/frr-k8s-jh6gj" Dec 06 03:47:38 crc kubenswrapper[4980]: I1206 03:47:38.415004 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/97b725b0-8d1d-4c85-8c96-f96e269da5c5-frr-sockets\") pod \"frr-k8s-jh6gj\" (UID: \"97b725b0-8d1d-4c85-8c96-f96e269da5c5\") " pod="metallb-system/frr-k8s-jh6gj" Dec 06 03:47:38 crc kubenswrapper[4980]: I1206 03:47:38.420979 4980 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"metallb-excludel2" Dec 06 03:47:38 crc kubenswrapper[4980]: I1206 03:47:38.421193 4980 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-certs-secret" Dec 06 03:47:38 crc kubenswrapper[4980]: I1206 03:47:38.421332 4980 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-dockercfg-7mvtw" Dec 06 03:47:38 crc kubenswrapper[4980]: I1206 03:47:38.429027 4980 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-memberlist" Dec 06 03:47:38 crc kubenswrapper[4980]: I1206 03:47:38.438246 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fx56t\" (UniqueName: \"kubernetes.io/projected/97b725b0-8d1d-4c85-8c96-f96e269da5c5-kube-api-access-fx56t\") pod \"frr-k8s-jh6gj\" (UID: \"97b725b0-8d1d-4c85-8c96-f96e269da5c5\") " pod="metallb-system/frr-k8s-jh6gj" Dec 06 03:47:38 crc kubenswrapper[4980]: I1206 03:47:38.443258 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z6g9k\" (UniqueName: \"kubernetes.io/projected/1f860039-c5ba-4ba9-90b1-ecb00110015f-kube-api-access-z6g9k\") pod \"frr-k8s-webhook-server-7fcb986d4-7qtqr\" (UID: \"1f860039-c5ba-4ba9-90b1-ecb00110015f\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-7qtqr" Dec 06 03:47:38 crc kubenswrapper[4980]: I1206 03:47:38.446590 4980 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/controller-f8648f98b-jg82p"] Dec 06 03:47:38 crc kubenswrapper[4980]: I1206 03:47:38.447668 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-f8648f98b-jg82p" Dec 06 03:47:38 crc kubenswrapper[4980]: I1206 03:47:38.450974 4980 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-certs-secret" Dec 06 03:47:38 crc kubenswrapper[4980]: I1206 03:47:38.451306 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/97b725b0-8d1d-4c85-8c96-f96e269da5c5-metrics-certs\") pod \"frr-k8s-jh6gj\" (UID: \"97b725b0-8d1d-4c85-8c96-f96e269da5c5\") " pod="metallb-system/frr-k8s-jh6gj" Dec 06 03:47:38 crc kubenswrapper[4980]: I1206 03:47:38.461478 4980 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-f8648f98b-jg82p"] Dec 06 03:47:38 crc kubenswrapper[4980]: I1206 03:47:38.513376 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n56sd\" (UniqueName: \"kubernetes.io/projected/53351890-4d11-40c9-91c9-b91c24578d70-kube-api-access-n56sd\") pod \"speaker-6wv7f\" (UID: \"53351890-4d11-40c9-91c9-b91c24578d70\") " pod="metallb-system/speaker-6wv7f" Dec 06 03:47:38 crc kubenswrapper[4980]: I1206 03:47:38.513458 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/53351890-4d11-40c9-91c9-b91c24578d70-metallb-excludel2\") pod \"speaker-6wv7f\" (UID: \"53351890-4d11-40c9-91c9-b91c24578d70\") " pod="metallb-system/speaker-6wv7f" Dec 06 03:47:38 crc kubenswrapper[4980]: I1206 03:47:38.513843 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/53351890-4d11-40c9-91c9-b91c24578d70-metrics-certs\") pod \"speaker-6wv7f\" (UID: \"53351890-4d11-40c9-91c9-b91c24578d70\") " pod="metallb-system/speaker-6wv7f" Dec 06 03:47:38 crc kubenswrapper[4980]: I1206 03:47:38.513947 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/53351890-4d11-40c9-91c9-b91c24578d70-memberlist\") pod \"speaker-6wv7f\" (UID: \"53351890-4d11-40c9-91c9-b91c24578d70\") " pod="metallb-system/speaker-6wv7f" Dec 06 03:47:38 crc kubenswrapper[4980]: I1206 03:47:38.605363 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-jh6gj" Dec 06 03:47:38 crc kubenswrapper[4980]: I1206 03:47:38.615234 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/a8cfafe9-920a-4bdd-a857-dae21346a492-cert\") pod \"controller-f8648f98b-jg82p\" (UID: \"a8cfafe9-920a-4bdd-a857-dae21346a492\") " pod="metallb-system/controller-f8648f98b-jg82p" Dec 06 03:47:38 crc kubenswrapper[4980]: I1206 03:47:38.615540 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/53351890-4d11-40c9-91c9-b91c24578d70-memberlist\") pod \"speaker-6wv7f\" (UID: \"53351890-4d11-40c9-91c9-b91c24578d70\") " pod="metallb-system/speaker-6wv7f" Dec 06 03:47:38 crc kubenswrapper[4980]: I1206 03:47:38.615590 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n56sd\" (UniqueName: \"kubernetes.io/projected/53351890-4d11-40c9-91c9-b91c24578d70-kube-api-access-n56sd\") pod \"speaker-6wv7f\" (UID: \"53351890-4d11-40c9-91c9-b91c24578d70\") " pod="metallb-system/speaker-6wv7f" Dec 06 03:47:38 crc kubenswrapper[4980]: I1206 03:47:38.615635 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2wkzz\" (UniqueName: \"kubernetes.io/projected/a8cfafe9-920a-4bdd-a857-dae21346a492-kube-api-access-2wkzz\") pod \"controller-f8648f98b-jg82p\" (UID: \"a8cfafe9-920a-4bdd-a857-dae21346a492\") " pod="metallb-system/controller-f8648f98b-jg82p" Dec 06 03:47:38 crc kubenswrapper[4980]: I1206 03:47:38.615662 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/53351890-4d11-40c9-91c9-b91c24578d70-metallb-excludel2\") pod \"speaker-6wv7f\" (UID: \"53351890-4d11-40c9-91c9-b91c24578d70\") " pod="metallb-system/speaker-6wv7f" Dec 06 03:47:38 crc kubenswrapper[4980]: I1206 03:47:38.615687 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/53351890-4d11-40c9-91c9-b91c24578d70-metrics-certs\") pod \"speaker-6wv7f\" (UID: \"53351890-4d11-40c9-91c9-b91c24578d70\") " pod="metallb-system/speaker-6wv7f" Dec 06 03:47:38 crc kubenswrapper[4980]: I1206 03:47:38.615730 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/a8cfafe9-920a-4bdd-a857-dae21346a492-metrics-certs\") pod \"controller-f8648f98b-jg82p\" (UID: \"a8cfafe9-920a-4bdd-a857-dae21346a492\") " pod="metallb-system/controller-f8648f98b-jg82p" Dec 06 03:47:38 crc kubenswrapper[4980]: E1206 03:47:38.615922 4980 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Dec 06 03:47:38 crc kubenswrapper[4980]: E1206 03:47:38.615977 4980 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/53351890-4d11-40c9-91c9-b91c24578d70-memberlist podName:53351890-4d11-40c9-91c9-b91c24578d70 nodeName:}" failed. No retries permitted until 2025-12-06 03:47:39.115962609 +0000 UTC m=+818.360713870 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/53351890-4d11-40c9-91c9-b91c24578d70-memberlist") pod "speaker-6wv7f" (UID: "53351890-4d11-40c9-91c9-b91c24578d70") : secret "metallb-memberlist" not found Dec 06 03:47:38 crc kubenswrapper[4980]: I1206 03:47:38.616867 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/53351890-4d11-40c9-91c9-b91c24578d70-metallb-excludel2\") pod \"speaker-6wv7f\" (UID: \"53351890-4d11-40c9-91c9-b91c24578d70\") " pod="metallb-system/speaker-6wv7f" Dec 06 03:47:38 crc kubenswrapper[4980]: I1206 03:47:38.622161 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/53351890-4d11-40c9-91c9-b91c24578d70-metrics-certs\") pod \"speaker-6wv7f\" (UID: \"53351890-4d11-40c9-91c9-b91c24578d70\") " pod="metallb-system/speaker-6wv7f" Dec 06 03:47:38 crc kubenswrapper[4980]: I1206 03:47:38.634918 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n56sd\" (UniqueName: \"kubernetes.io/projected/53351890-4d11-40c9-91c9-b91c24578d70-kube-api-access-n56sd\") pod \"speaker-6wv7f\" (UID: \"53351890-4d11-40c9-91c9-b91c24578d70\") " pod="metallb-system/speaker-6wv7f" Dec 06 03:47:38 crc kubenswrapper[4980]: I1206 03:47:38.716613 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/a8cfafe9-920a-4bdd-a857-dae21346a492-cert\") pod \"controller-f8648f98b-jg82p\" (UID: \"a8cfafe9-920a-4bdd-a857-dae21346a492\") " pod="metallb-system/controller-f8648f98b-jg82p" Dec 06 03:47:38 crc kubenswrapper[4980]: I1206 03:47:38.716718 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2wkzz\" (UniqueName: \"kubernetes.io/projected/a8cfafe9-920a-4bdd-a857-dae21346a492-kube-api-access-2wkzz\") pod \"controller-f8648f98b-jg82p\" (UID: \"a8cfafe9-920a-4bdd-a857-dae21346a492\") " pod="metallb-system/controller-f8648f98b-jg82p" Dec 06 03:47:38 crc kubenswrapper[4980]: I1206 03:47:38.716769 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/a8cfafe9-920a-4bdd-a857-dae21346a492-metrics-certs\") pod \"controller-f8648f98b-jg82p\" (UID: \"a8cfafe9-920a-4bdd-a857-dae21346a492\") " pod="metallb-system/controller-f8648f98b-jg82p" Dec 06 03:47:38 crc kubenswrapper[4980]: I1206 03:47:38.718749 4980 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Dec 06 03:47:38 crc kubenswrapper[4980]: I1206 03:47:38.720454 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/a8cfafe9-920a-4bdd-a857-dae21346a492-metrics-certs\") pod \"controller-f8648f98b-jg82p\" (UID: \"a8cfafe9-920a-4bdd-a857-dae21346a492\") " pod="metallb-system/controller-f8648f98b-jg82p" Dec 06 03:47:38 crc kubenswrapper[4980]: I1206 03:47:38.731598 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/a8cfafe9-920a-4bdd-a857-dae21346a492-cert\") pod \"controller-f8648f98b-jg82p\" (UID: \"a8cfafe9-920a-4bdd-a857-dae21346a492\") " pod="metallb-system/controller-f8648f98b-jg82p" Dec 06 03:47:38 crc kubenswrapper[4980]: I1206 03:47:38.733255 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2wkzz\" (UniqueName: \"kubernetes.io/projected/a8cfafe9-920a-4bdd-a857-dae21346a492-kube-api-access-2wkzz\") pod \"controller-f8648f98b-jg82p\" (UID: \"a8cfafe9-920a-4bdd-a857-dae21346a492\") " pod="metallb-system/controller-f8648f98b-jg82p" Dec 06 03:47:38 crc kubenswrapper[4980]: I1206 03:47:38.783284 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-f8648f98b-jg82p" Dec 06 03:47:38 crc kubenswrapper[4980]: I1206 03:47:38.919581 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/1f860039-c5ba-4ba9-90b1-ecb00110015f-cert\") pod \"frr-k8s-webhook-server-7fcb986d4-7qtqr\" (UID: \"1f860039-c5ba-4ba9-90b1-ecb00110015f\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-7qtqr" Dec 06 03:47:38 crc kubenswrapper[4980]: E1206 03:47:38.919791 4980 secret.go:188] Couldn't get secret metallb-system/frr-k8s-webhook-server-cert: secret "frr-k8s-webhook-server-cert" not found Dec 06 03:47:38 crc kubenswrapper[4980]: E1206 03:47:38.919872 4980 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/1f860039-c5ba-4ba9-90b1-ecb00110015f-cert podName:1f860039-c5ba-4ba9-90b1-ecb00110015f nodeName:}" failed. No retries permitted until 2025-12-06 03:47:39.919853908 +0000 UTC m=+819.164605169 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/1f860039-c5ba-4ba9-90b1-ecb00110015f-cert") pod "frr-k8s-webhook-server-7fcb986d4-7qtqr" (UID: "1f860039-c5ba-4ba9-90b1-ecb00110015f") : secret "frr-k8s-webhook-server-cert" not found Dec 06 03:47:39 crc kubenswrapper[4980]: I1206 03:47:39.122860 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/53351890-4d11-40c9-91c9-b91c24578d70-memberlist\") pod \"speaker-6wv7f\" (UID: \"53351890-4d11-40c9-91c9-b91c24578d70\") " pod="metallb-system/speaker-6wv7f" Dec 06 03:47:39 crc kubenswrapper[4980]: E1206 03:47:39.123146 4980 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Dec 06 03:47:39 crc kubenswrapper[4980]: E1206 03:47:39.123293 4980 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/53351890-4d11-40c9-91c9-b91c24578d70-memberlist podName:53351890-4d11-40c9-91c9-b91c24578d70 nodeName:}" failed. No retries permitted until 2025-12-06 03:47:40.123251874 +0000 UTC m=+819.368003185 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/53351890-4d11-40c9-91c9-b91c24578d70-memberlist") pod "speaker-6wv7f" (UID: "53351890-4d11-40c9-91c9-b91c24578d70") : secret "metallb-memberlist" not found Dec 06 03:47:39 crc kubenswrapper[4980]: I1206 03:47:39.217977 4980 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-f8648f98b-jg82p"] Dec 06 03:47:39 crc kubenswrapper[4980]: W1206 03:47:39.228124 4980 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda8cfafe9_920a_4bdd_a857_dae21346a492.slice/crio-be707b348876f3b2b42f87aa2484e1fe02d0216e952e1955e8eab98e02388e54 WatchSource:0}: Error finding container be707b348876f3b2b42f87aa2484e1fe02d0216e952e1955e8eab98e02388e54: Status 404 returned error can't find the container with id be707b348876f3b2b42f87aa2484e1fe02d0216e952e1955e8eab98e02388e54 Dec 06 03:47:39 crc kubenswrapper[4980]: I1206 03:47:39.506735 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-f8648f98b-jg82p" event={"ID":"a8cfafe9-920a-4bdd-a857-dae21346a492","Type":"ContainerStarted","Data":"be707b348876f3b2b42f87aa2484e1fe02d0216e952e1955e8eab98e02388e54"} Dec 06 03:47:39 crc kubenswrapper[4980]: I1206 03:47:39.508747 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-jh6gj" event={"ID":"97b725b0-8d1d-4c85-8c96-f96e269da5c5","Type":"ContainerStarted","Data":"a6f9c06729ee0b78be6277a66751755f73ef1af527d6779e9356c078790ac103"} Dec 06 03:47:39 crc kubenswrapper[4980]: I1206 03:47:39.935691 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/1f860039-c5ba-4ba9-90b1-ecb00110015f-cert\") pod \"frr-k8s-webhook-server-7fcb986d4-7qtqr\" (UID: \"1f860039-c5ba-4ba9-90b1-ecb00110015f\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-7qtqr" Dec 06 03:47:39 crc kubenswrapper[4980]: I1206 03:47:39.947381 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/1f860039-c5ba-4ba9-90b1-ecb00110015f-cert\") pod \"frr-k8s-webhook-server-7fcb986d4-7qtqr\" (UID: \"1f860039-c5ba-4ba9-90b1-ecb00110015f\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-7qtqr" Dec 06 03:47:40 crc kubenswrapper[4980]: I1206 03:47:40.114374 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-7qtqr" Dec 06 03:47:40 crc kubenswrapper[4980]: I1206 03:47:40.139297 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/53351890-4d11-40c9-91c9-b91c24578d70-memberlist\") pod \"speaker-6wv7f\" (UID: \"53351890-4d11-40c9-91c9-b91c24578d70\") " pod="metallb-system/speaker-6wv7f" Dec 06 03:47:40 crc kubenswrapper[4980]: E1206 03:47:40.139616 4980 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Dec 06 03:47:40 crc kubenswrapper[4980]: E1206 03:47:40.139757 4980 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/53351890-4d11-40c9-91c9-b91c24578d70-memberlist podName:53351890-4d11-40c9-91c9-b91c24578d70 nodeName:}" failed. No retries permitted until 2025-12-06 03:47:42.139723569 +0000 UTC m=+821.384474880 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/53351890-4d11-40c9-91c9-b91c24578d70-memberlist") pod "speaker-6wv7f" (UID: "53351890-4d11-40c9-91c9-b91c24578d70") : secret "metallb-memberlist" not found Dec 06 03:47:40 crc kubenswrapper[4980]: I1206 03:47:40.386646 4980 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-7fcb986d4-7qtqr"] Dec 06 03:47:40 crc kubenswrapper[4980]: I1206 03:47:40.518360 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-f8648f98b-jg82p" event={"ID":"a8cfafe9-920a-4bdd-a857-dae21346a492","Type":"ContainerStarted","Data":"e862a70684693b4b863c939e9d35e6e039836715d4d146a162a616a8bae232e4"} Dec 06 03:47:40 crc kubenswrapper[4980]: I1206 03:47:40.519800 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-7qtqr" event={"ID":"1f860039-c5ba-4ba9-90b1-ecb00110015f","Type":"ContainerStarted","Data":"ec531167e6f0a26fe3af9fa6929e93dffa966d962ec9fbf38ec3a5b94923edc2"} Dec 06 03:47:42 crc kubenswrapper[4980]: I1206 03:47:42.175665 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/53351890-4d11-40c9-91c9-b91c24578d70-memberlist\") pod \"speaker-6wv7f\" (UID: \"53351890-4d11-40c9-91c9-b91c24578d70\") " pod="metallb-system/speaker-6wv7f" Dec 06 03:47:42 crc kubenswrapper[4980]: I1206 03:47:42.190259 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/53351890-4d11-40c9-91c9-b91c24578d70-memberlist\") pod \"speaker-6wv7f\" (UID: \"53351890-4d11-40c9-91c9-b91c24578d70\") " pod="metallb-system/speaker-6wv7f" Dec 06 03:47:42 crc kubenswrapper[4980]: I1206 03:47:42.372286 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-6wv7f" Dec 06 03:47:42 crc kubenswrapper[4980]: W1206 03:47:42.399881 4980 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod53351890_4d11_40c9_91c9_b91c24578d70.slice/crio-36102e1bcf4cde41bff6c655cc1d856d0d85a560b3903cf711d9acbfd76cd12a WatchSource:0}: Error finding container 36102e1bcf4cde41bff6c655cc1d856d0d85a560b3903cf711d9acbfd76cd12a: Status 404 returned error can't find the container with id 36102e1bcf4cde41bff6c655cc1d856d0d85a560b3903cf711d9acbfd76cd12a Dec 06 03:47:42 crc kubenswrapper[4980]: I1206 03:47:42.568922 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-6wv7f" event={"ID":"53351890-4d11-40c9-91c9-b91c24578d70","Type":"ContainerStarted","Data":"36102e1bcf4cde41bff6c655cc1d856d0d85a560b3903cf711d9acbfd76cd12a"} Dec 06 03:47:43 crc kubenswrapper[4980]: I1206 03:47:43.580007 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-6wv7f" event={"ID":"53351890-4d11-40c9-91c9-b91c24578d70","Type":"ContainerStarted","Data":"924fef560bbdc94de402a32e010f42b485a93859c965d1175704fb320d9407fc"} Dec 06 03:47:49 crc kubenswrapper[4980]: I1206 03:47:49.620410 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-f8648f98b-jg82p" event={"ID":"a8cfafe9-920a-4bdd-a857-dae21346a492","Type":"ContainerStarted","Data":"97cf2217c50da1dd4709193d148ca2b40c3a92a5d17b5c49340eae610a2f63c1"} Dec 06 03:47:49 crc kubenswrapper[4980]: I1206 03:47:49.621007 4980 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/controller-f8648f98b-jg82p" Dec 06 03:47:49 crc kubenswrapper[4980]: I1206 03:47:49.622104 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-7qtqr" event={"ID":"1f860039-c5ba-4ba9-90b1-ecb00110015f","Type":"ContainerStarted","Data":"bfb429f9888b8158eb6dc1c8537c9eb0b77835a042ebfdaec134c0fafc0e2c77"} Dec 06 03:47:49 crc kubenswrapper[4980]: I1206 03:47:49.622222 4980 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-7qtqr" Dec 06 03:47:49 crc kubenswrapper[4980]: I1206 03:47:49.623988 4980 generic.go:334] "Generic (PLEG): container finished" podID="97b725b0-8d1d-4c85-8c96-f96e269da5c5" containerID="605c263b94538c8fbed4afa2dd6c2d547fba22e3f201515029b3ff3db04141b2" exitCode=0 Dec 06 03:47:49 crc kubenswrapper[4980]: I1206 03:47:49.624042 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-jh6gj" event={"ID":"97b725b0-8d1d-4c85-8c96-f96e269da5c5","Type":"ContainerDied","Data":"605c263b94538c8fbed4afa2dd6c2d547fba22e3f201515029b3ff3db04141b2"} Dec 06 03:47:49 crc kubenswrapper[4980]: I1206 03:47:49.626215 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-6wv7f" event={"ID":"53351890-4d11-40c9-91c9-b91c24578d70","Type":"ContainerStarted","Data":"08912e18a7db3eec93797f7832df0c6f32606ea5b8ceaa8edf970f67a924c5c9"} Dec 06 03:47:49 crc kubenswrapper[4980]: I1206 03:47:49.626380 4980 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/speaker-6wv7f" Dec 06 03:47:49 crc kubenswrapper[4980]: I1206 03:47:49.663391 4980 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/controller-f8648f98b-jg82p" podStartSLOduration=3.145806714 podStartE2EDuration="11.663361074s" podCreationTimestamp="2025-12-06 03:47:38 +0000 UTC" firstStartedPulling="2025-12-06 03:47:40.367827959 +0000 UTC m=+819.612579220" lastFinishedPulling="2025-12-06 03:47:48.885382299 +0000 UTC m=+828.130133580" observedRunningTime="2025-12-06 03:47:49.646818016 +0000 UTC m=+828.891569327" watchObservedRunningTime="2025-12-06 03:47:49.663361074 +0000 UTC m=+828.908112335" Dec 06 03:47:49 crc kubenswrapper[4980]: I1206 03:47:49.681747 4980 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/speaker-6wv7f" podStartSLOduration=5.600835846 podStartE2EDuration="11.681731725s" podCreationTimestamp="2025-12-06 03:47:38 +0000 UTC" firstStartedPulling="2025-12-06 03:47:42.807376864 +0000 UTC m=+822.052128135" lastFinishedPulling="2025-12-06 03:47:48.888272743 +0000 UTC m=+828.133024014" observedRunningTime="2025-12-06 03:47:49.678309586 +0000 UTC m=+828.923060867" watchObservedRunningTime="2025-12-06 03:47:49.681731725 +0000 UTC m=+828.926482986" Dec 06 03:47:49 crc kubenswrapper[4980]: I1206 03:47:49.698269 4980 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-7qtqr" podStartSLOduration=3.092495503 podStartE2EDuration="11.698251072s" podCreationTimestamp="2025-12-06 03:47:38 +0000 UTC" firstStartedPulling="2025-12-06 03:47:40.414577729 +0000 UTC m=+819.659329000" lastFinishedPulling="2025-12-06 03:47:49.020333298 +0000 UTC m=+828.265084569" observedRunningTime="2025-12-06 03:47:49.697635874 +0000 UTC m=+828.942387185" watchObservedRunningTime="2025-12-06 03:47:49.698251072 +0000 UTC m=+828.943002343" Dec 06 03:47:50 crc kubenswrapper[4980]: I1206 03:47:50.643762 4980 generic.go:334] "Generic (PLEG): container finished" podID="97b725b0-8d1d-4c85-8c96-f96e269da5c5" containerID="3f6b627c92b25325d53364e48a8535a9993e391a7fc0109661f5399d6ce4be8c" exitCode=0 Dec 06 03:47:50 crc kubenswrapper[4980]: I1206 03:47:50.643919 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-jh6gj" event={"ID":"97b725b0-8d1d-4c85-8c96-f96e269da5c5","Type":"ContainerDied","Data":"3f6b627c92b25325d53364e48a8535a9993e391a7fc0109661f5399d6ce4be8c"} Dec 06 03:47:50 crc kubenswrapper[4980]: I1206 03:47:50.649425 4980 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/controller-f8648f98b-jg82p" Dec 06 03:47:51 crc kubenswrapper[4980]: I1206 03:47:51.651807 4980 generic.go:334] "Generic (PLEG): container finished" podID="97b725b0-8d1d-4c85-8c96-f96e269da5c5" containerID="84fe3f7337182774cbf9f26e74e68f6ee377d28021f5fa14ac6bbb98f1f67c96" exitCode=0 Dec 06 03:47:51 crc kubenswrapper[4980]: I1206 03:47:51.651885 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-jh6gj" event={"ID":"97b725b0-8d1d-4c85-8c96-f96e269da5c5","Type":"ContainerDied","Data":"84fe3f7337182774cbf9f26e74e68f6ee377d28021f5fa14ac6bbb98f1f67c96"} Dec 06 03:47:52 crc kubenswrapper[4980]: I1206 03:47:52.376431 4980 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/speaker-6wv7f" Dec 06 03:47:52 crc kubenswrapper[4980]: I1206 03:47:52.694642 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-jh6gj" event={"ID":"97b725b0-8d1d-4c85-8c96-f96e269da5c5","Type":"ContainerStarted","Data":"1098c41c24162be099865ba8926c9800f903f18fa0b3ad0177db2b40d4cc9fc6"} Dec 06 03:47:52 crc kubenswrapper[4980]: I1206 03:47:52.694955 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-jh6gj" event={"ID":"97b725b0-8d1d-4c85-8c96-f96e269da5c5","Type":"ContainerStarted","Data":"058fc9ac934481fe0f7b978ba4fd53bb65be0e93051ead03def742981bfd61dc"} Dec 06 03:47:52 crc kubenswrapper[4980]: I1206 03:47:52.694969 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-jh6gj" event={"ID":"97b725b0-8d1d-4c85-8c96-f96e269da5c5","Type":"ContainerStarted","Data":"810d20ade001b84422932592ecb7a8bb74a07119106c2f3dbeda564476c68678"} Dec 06 03:47:52 crc kubenswrapper[4980]: I1206 03:47:52.694980 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-jh6gj" event={"ID":"97b725b0-8d1d-4c85-8c96-f96e269da5c5","Type":"ContainerStarted","Data":"1684731032bc04bcfe23596f5f7778150f1f7648f2ec943dc25c82ebe2e8af5a"} Dec 06 03:47:52 crc kubenswrapper[4980]: I1206 03:47:52.694992 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-jh6gj" event={"ID":"97b725b0-8d1d-4c85-8c96-f96e269da5c5","Type":"ContainerStarted","Data":"2641a49088eced0e8e320eb041db8851bebd553cbd851b58ebd8868187042274"} Dec 06 03:47:53 crc kubenswrapper[4980]: I1206 03:47:53.708764 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-jh6gj" event={"ID":"97b725b0-8d1d-4c85-8c96-f96e269da5c5","Type":"ContainerStarted","Data":"d908762619cd9808a940eebc545489531af4a07b11f9f9cef1aecf4e1e41b0da"} Dec 06 03:47:53 crc kubenswrapper[4980]: I1206 03:47:53.708997 4980 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-jh6gj" Dec 06 03:47:53 crc kubenswrapper[4980]: I1206 03:47:53.735674 4980 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-jh6gj" podStartSLOduration=5.436796577 podStartE2EDuration="15.735632097s" podCreationTimestamp="2025-12-06 03:47:38 +0000 UTC" firstStartedPulling="2025-12-06 03:47:38.725573276 +0000 UTC m=+817.970324557" lastFinishedPulling="2025-12-06 03:47:49.024408806 +0000 UTC m=+828.269160077" observedRunningTime="2025-12-06 03:47:53.730070736 +0000 UTC m=+832.974822007" watchObservedRunningTime="2025-12-06 03:47:53.735632097 +0000 UTC m=+832.980383378" Dec 06 03:47:58 crc kubenswrapper[4980]: I1206 03:47:58.605273 4980 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="metallb-system/frr-k8s-jh6gj" Dec 06 03:47:58 crc kubenswrapper[4980]: I1206 03:47:58.658176 4980 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="metallb-system/frr-k8s-jh6gj" Dec 06 03:47:59 crc kubenswrapper[4980]: I1206 03:47:59.093604 4980 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/mariadb-operator-index-6vs7m"] Dec 06 03:47:59 crc kubenswrapper[4980]: I1206 03:47:59.094618 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-index-6vs7m" Dec 06 03:47:59 crc kubenswrapper[4980]: I1206 03:47:59.099315 4980 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"mariadb-operator-index-dockercfg-smcds" Dec 06 03:47:59 crc kubenswrapper[4980]: I1206 03:47:59.100153 4980 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"openshift-service-ca.crt" Dec 06 03:47:59 crc kubenswrapper[4980]: I1206 03:47:59.107148 4980 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"kube-root-ca.crt" Dec 06 03:47:59 crc kubenswrapper[4980]: I1206 03:47:59.126666 4980 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-index-6vs7m"] Dec 06 03:47:59 crc kubenswrapper[4980]: I1206 03:47:59.283587 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5hzxl\" (UniqueName: \"kubernetes.io/projected/8893758a-10c8-416d-a93e-db4c41980819-kube-api-access-5hzxl\") pod \"mariadb-operator-index-6vs7m\" (UID: \"8893758a-10c8-416d-a93e-db4c41980819\") " pod="openstack-operators/mariadb-operator-index-6vs7m" Dec 06 03:47:59 crc kubenswrapper[4980]: I1206 03:47:59.385310 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5hzxl\" (UniqueName: \"kubernetes.io/projected/8893758a-10c8-416d-a93e-db4c41980819-kube-api-access-5hzxl\") pod \"mariadb-operator-index-6vs7m\" (UID: \"8893758a-10c8-416d-a93e-db4c41980819\") " pod="openstack-operators/mariadb-operator-index-6vs7m" Dec 06 03:47:59 crc kubenswrapper[4980]: I1206 03:47:59.406078 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5hzxl\" (UniqueName: \"kubernetes.io/projected/8893758a-10c8-416d-a93e-db4c41980819-kube-api-access-5hzxl\") pod \"mariadb-operator-index-6vs7m\" (UID: \"8893758a-10c8-416d-a93e-db4c41980819\") " pod="openstack-operators/mariadb-operator-index-6vs7m" Dec 06 03:47:59 crc kubenswrapper[4980]: I1206 03:47:59.413599 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-index-6vs7m" Dec 06 03:48:00 crc kubenswrapper[4980]: I1206 03:48:00.118383 4980 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-7qtqr" Dec 06 03:48:00 crc kubenswrapper[4980]: I1206 03:48:00.356442 4980 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-index-6vs7m"] Dec 06 03:48:00 crc kubenswrapper[4980]: W1206 03:48:00.362740 4980 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8893758a_10c8_416d_a93e_db4c41980819.slice/crio-c728a9bfc57156f88cc45ab87da4de49d0020de6a0df74c77e972e85f64c7c9f WatchSource:0}: Error finding container c728a9bfc57156f88cc45ab87da4de49d0020de6a0df74c77e972e85f64c7c9f: Status 404 returned error can't find the container with id c728a9bfc57156f88cc45ab87da4de49d0020de6a0df74c77e972e85f64c7c9f Dec 06 03:48:00 crc kubenswrapper[4980]: I1206 03:48:00.761623 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-index-6vs7m" event={"ID":"8893758a-10c8-416d-a93e-db4c41980819","Type":"ContainerStarted","Data":"c728a9bfc57156f88cc45ab87da4de49d0020de6a0df74c77e972e85f64c7c9f"} Dec 06 03:48:02 crc kubenswrapper[4980]: I1206 03:48:02.417738 4980 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/mariadb-operator-index-6vs7m"] Dec 06 03:48:03 crc kubenswrapper[4980]: I1206 03:48:03.021791 4980 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/mariadb-operator-index-5wm7n"] Dec 06 03:48:03 crc kubenswrapper[4980]: I1206 03:48:03.022624 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-index-5wm7n" Dec 06 03:48:03 crc kubenswrapper[4980]: I1206 03:48:03.034187 4980 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-index-5wm7n"] Dec 06 03:48:03 crc kubenswrapper[4980]: I1206 03:48:03.135749 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j7vrc\" (UniqueName: \"kubernetes.io/projected/4066d752-9e8f-47d6-a553-f7f7e7c06e25-kube-api-access-j7vrc\") pod \"mariadb-operator-index-5wm7n\" (UID: \"4066d752-9e8f-47d6-a553-f7f7e7c06e25\") " pod="openstack-operators/mariadb-operator-index-5wm7n" Dec 06 03:48:03 crc kubenswrapper[4980]: I1206 03:48:03.237905 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j7vrc\" (UniqueName: \"kubernetes.io/projected/4066d752-9e8f-47d6-a553-f7f7e7c06e25-kube-api-access-j7vrc\") pod \"mariadb-operator-index-5wm7n\" (UID: \"4066d752-9e8f-47d6-a553-f7f7e7c06e25\") " pod="openstack-operators/mariadb-operator-index-5wm7n" Dec 06 03:48:03 crc kubenswrapper[4980]: I1206 03:48:03.256122 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j7vrc\" (UniqueName: \"kubernetes.io/projected/4066d752-9e8f-47d6-a553-f7f7e7c06e25-kube-api-access-j7vrc\") pod \"mariadb-operator-index-5wm7n\" (UID: \"4066d752-9e8f-47d6-a553-f7f7e7c06e25\") " pod="openstack-operators/mariadb-operator-index-5wm7n" Dec 06 03:48:03 crc kubenswrapper[4980]: I1206 03:48:03.344790 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-index-5wm7n" Dec 06 03:48:03 crc kubenswrapper[4980]: I1206 03:48:03.880800 4980 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-index-5wm7n"] Dec 06 03:48:08 crc kubenswrapper[4980]: I1206 03:48:08.612886 4980 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-jh6gj" Dec 06 03:48:12 crc kubenswrapper[4980]: W1206 03:48:12.594595 4980 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4066d752_9e8f_47d6_a553_f7f7e7c06e25.slice/crio-75c254253f519ed2210876033bd264828e652d1ef71144189dc66e9b40f04896 WatchSource:0}: Error finding container 75c254253f519ed2210876033bd264828e652d1ef71144189dc66e9b40f04896: Status 404 returned error can't find the container with id 75c254253f519ed2210876033bd264828e652d1ef71144189dc66e9b40f04896 Dec 06 03:48:12 crc kubenswrapper[4980]: I1206 03:48:12.844572 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-index-5wm7n" event={"ID":"4066d752-9e8f-47d6-a553-f7f7e7c06e25","Type":"ContainerStarted","Data":"75c254253f519ed2210876033bd264828e652d1ef71144189dc66e9b40f04896"} Dec 06 03:48:13 crc kubenswrapper[4980]: I1206 03:48:13.853660 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-index-5wm7n" event={"ID":"4066d752-9e8f-47d6-a553-f7f7e7c06e25","Type":"ContainerStarted","Data":"829b2e6897b975958c7de225560322681da6b2c5c919a702006282ba1112f1bf"} Dec 06 03:48:13 crc kubenswrapper[4980]: I1206 03:48:13.856381 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-index-6vs7m" event={"ID":"8893758a-10c8-416d-a93e-db4c41980819","Type":"ContainerStarted","Data":"88540a4f4e6808f31d45a0bd5c5c88604e135f7c6a240a9de42f9a3421327419"} Dec 06 03:48:13 crc kubenswrapper[4980]: I1206 03:48:13.856532 4980 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/mariadb-operator-index-6vs7m" podUID="8893758a-10c8-416d-a93e-db4c41980819" containerName="registry-server" containerID="cri-o://88540a4f4e6808f31d45a0bd5c5c88604e135f7c6a240a9de42f9a3421327419" gracePeriod=2 Dec 06 03:48:13 crc kubenswrapper[4980]: I1206 03:48:13.879536 4980 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/mariadb-operator-index-5wm7n" podStartSLOduration=10.485102986 podStartE2EDuration="10.8794954s" podCreationTimestamp="2025-12-06 03:48:03 +0000 UTC" firstStartedPulling="2025-12-06 03:48:12.596972869 +0000 UTC m=+851.841724160" lastFinishedPulling="2025-12-06 03:48:12.991365273 +0000 UTC m=+852.236116574" observedRunningTime="2025-12-06 03:48:13.87883022 +0000 UTC m=+853.123581491" watchObservedRunningTime="2025-12-06 03:48:13.8794954 +0000 UTC m=+853.124246661" Dec 06 03:48:13 crc kubenswrapper[4980]: I1206 03:48:13.901036 4980 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/mariadb-operator-index-6vs7m" podStartSLOduration=2.4159072200000002 podStartE2EDuration="14.901017381s" podCreationTimestamp="2025-12-06 03:47:59 +0000 UTC" firstStartedPulling="2025-12-06 03:48:00.364338692 +0000 UTC m=+839.609089973" lastFinishedPulling="2025-12-06 03:48:12.849448863 +0000 UTC m=+852.094200134" observedRunningTime="2025-12-06 03:48:13.899500148 +0000 UTC m=+853.144251449" watchObservedRunningTime="2025-12-06 03:48:13.901017381 +0000 UTC m=+853.145768652" Dec 06 03:48:14 crc kubenswrapper[4980]: I1206 03:48:14.276372 4980 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-index-6vs7m" Dec 06 03:48:14 crc kubenswrapper[4980]: I1206 03:48:14.437746 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5hzxl\" (UniqueName: \"kubernetes.io/projected/8893758a-10c8-416d-a93e-db4c41980819-kube-api-access-5hzxl\") pod \"8893758a-10c8-416d-a93e-db4c41980819\" (UID: \"8893758a-10c8-416d-a93e-db4c41980819\") " Dec 06 03:48:14 crc kubenswrapper[4980]: I1206 03:48:14.442795 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8893758a-10c8-416d-a93e-db4c41980819-kube-api-access-5hzxl" (OuterVolumeSpecName: "kube-api-access-5hzxl") pod "8893758a-10c8-416d-a93e-db4c41980819" (UID: "8893758a-10c8-416d-a93e-db4c41980819"). InnerVolumeSpecName "kube-api-access-5hzxl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 03:48:14 crc kubenswrapper[4980]: I1206 03:48:14.539206 4980 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5hzxl\" (UniqueName: \"kubernetes.io/projected/8893758a-10c8-416d-a93e-db4c41980819-kube-api-access-5hzxl\") on node \"crc\" DevicePath \"\"" Dec 06 03:48:14 crc kubenswrapper[4980]: I1206 03:48:14.867559 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-index-6vs7m" event={"ID":"8893758a-10c8-416d-a93e-db4c41980819","Type":"ContainerDied","Data":"88540a4f4e6808f31d45a0bd5c5c88604e135f7c6a240a9de42f9a3421327419"} Dec 06 03:48:14 crc kubenswrapper[4980]: I1206 03:48:14.867552 4980 generic.go:334] "Generic (PLEG): container finished" podID="8893758a-10c8-416d-a93e-db4c41980819" containerID="88540a4f4e6808f31d45a0bd5c5c88604e135f7c6a240a9de42f9a3421327419" exitCode=0 Dec 06 03:48:14 crc kubenswrapper[4980]: I1206 03:48:14.867652 4980 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-index-6vs7m" Dec 06 03:48:14 crc kubenswrapper[4980]: I1206 03:48:14.867669 4980 scope.go:117] "RemoveContainer" containerID="88540a4f4e6808f31d45a0bd5c5c88604e135f7c6a240a9de42f9a3421327419" Dec 06 03:48:14 crc kubenswrapper[4980]: I1206 03:48:14.867726 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-index-6vs7m" event={"ID":"8893758a-10c8-416d-a93e-db4c41980819","Type":"ContainerDied","Data":"c728a9bfc57156f88cc45ab87da4de49d0020de6a0df74c77e972e85f64c7c9f"} Dec 06 03:48:14 crc kubenswrapper[4980]: I1206 03:48:14.890603 4980 scope.go:117] "RemoveContainer" containerID="88540a4f4e6808f31d45a0bd5c5c88604e135f7c6a240a9de42f9a3421327419" Dec 06 03:48:14 crc kubenswrapper[4980]: E1206 03:48:14.891334 4980 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"88540a4f4e6808f31d45a0bd5c5c88604e135f7c6a240a9de42f9a3421327419\": container with ID starting with 88540a4f4e6808f31d45a0bd5c5c88604e135f7c6a240a9de42f9a3421327419 not found: ID does not exist" containerID="88540a4f4e6808f31d45a0bd5c5c88604e135f7c6a240a9de42f9a3421327419" Dec 06 03:48:14 crc kubenswrapper[4980]: I1206 03:48:14.891394 4980 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"88540a4f4e6808f31d45a0bd5c5c88604e135f7c6a240a9de42f9a3421327419"} err="failed to get container status \"88540a4f4e6808f31d45a0bd5c5c88604e135f7c6a240a9de42f9a3421327419\": rpc error: code = NotFound desc = could not find container \"88540a4f4e6808f31d45a0bd5c5c88604e135f7c6a240a9de42f9a3421327419\": container with ID starting with 88540a4f4e6808f31d45a0bd5c5c88604e135f7c6a240a9de42f9a3421327419 not found: ID does not exist" Dec 06 03:48:14 crc kubenswrapper[4980]: I1206 03:48:14.922998 4980 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/mariadb-operator-index-6vs7m"] Dec 06 03:48:14 crc kubenswrapper[4980]: I1206 03:48:14.935387 4980 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/mariadb-operator-index-6vs7m"] Dec 06 03:48:15 crc kubenswrapper[4980]: I1206 03:48:15.195326 4980 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8893758a-10c8-416d-a93e-db4c41980819" path="/var/lib/kubelet/pods/8893758a-10c8-416d-a93e-db4c41980819/volumes" Dec 06 03:48:23 crc kubenswrapper[4980]: I1206 03:48:23.346179 4980 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/mariadb-operator-index-5wm7n" Dec 06 03:48:23 crc kubenswrapper[4980]: I1206 03:48:23.346742 4980 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/mariadb-operator-index-5wm7n" Dec 06 03:48:23 crc kubenswrapper[4980]: I1206 03:48:23.389318 4980 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/mariadb-operator-index-5wm7n" Dec 06 03:48:23 crc kubenswrapper[4980]: I1206 03:48:23.970084 4980 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/mariadb-operator-index-5wm7n" Dec 06 03:48:30 crc kubenswrapper[4980]: I1206 03:48:30.834023 4980 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/798531b9d9a078af26f5f153dd8093f0980ac32bb052a41050c010ef74b88wp"] Dec 06 03:48:30 crc kubenswrapper[4980]: E1206 03:48:30.834707 4980 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8893758a-10c8-416d-a93e-db4c41980819" containerName="registry-server" Dec 06 03:48:30 crc kubenswrapper[4980]: I1206 03:48:30.834719 4980 state_mem.go:107] "Deleted CPUSet assignment" podUID="8893758a-10c8-416d-a93e-db4c41980819" containerName="registry-server" Dec 06 03:48:30 crc kubenswrapper[4980]: I1206 03:48:30.834840 4980 memory_manager.go:354] "RemoveStaleState removing state" podUID="8893758a-10c8-416d-a93e-db4c41980819" containerName="registry-server" Dec 06 03:48:30 crc kubenswrapper[4980]: I1206 03:48:30.835657 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/798531b9d9a078af26f5f153dd8093f0980ac32bb052a41050c010ef74b88wp" Dec 06 03:48:30 crc kubenswrapper[4980]: I1206 03:48:30.837376 4980 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"default-dockercfg-brs8h" Dec 06 03:48:30 crc kubenswrapper[4980]: I1206 03:48:30.845122 4980 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/798531b9d9a078af26f5f153dd8093f0980ac32bb052a41050c010ef74b88wp"] Dec 06 03:48:31 crc kubenswrapper[4980]: I1206 03:48:31.015724 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k8thz\" (UniqueName: \"kubernetes.io/projected/33b3d1b9-c7e6-44e3-9259-1251e4e340a6-kube-api-access-k8thz\") pod \"798531b9d9a078af26f5f153dd8093f0980ac32bb052a41050c010ef74b88wp\" (UID: \"33b3d1b9-c7e6-44e3-9259-1251e4e340a6\") " pod="openstack-operators/798531b9d9a078af26f5f153dd8093f0980ac32bb052a41050c010ef74b88wp" Dec 06 03:48:31 crc kubenswrapper[4980]: I1206 03:48:31.015783 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/33b3d1b9-c7e6-44e3-9259-1251e4e340a6-bundle\") pod \"798531b9d9a078af26f5f153dd8093f0980ac32bb052a41050c010ef74b88wp\" (UID: \"33b3d1b9-c7e6-44e3-9259-1251e4e340a6\") " pod="openstack-operators/798531b9d9a078af26f5f153dd8093f0980ac32bb052a41050c010ef74b88wp" Dec 06 03:48:31 crc kubenswrapper[4980]: I1206 03:48:31.015870 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/33b3d1b9-c7e6-44e3-9259-1251e4e340a6-util\") pod \"798531b9d9a078af26f5f153dd8093f0980ac32bb052a41050c010ef74b88wp\" (UID: \"33b3d1b9-c7e6-44e3-9259-1251e4e340a6\") " pod="openstack-operators/798531b9d9a078af26f5f153dd8093f0980ac32bb052a41050c010ef74b88wp" Dec 06 03:48:31 crc kubenswrapper[4980]: I1206 03:48:31.117476 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/33b3d1b9-c7e6-44e3-9259-1251e4e340a6-bundle\") pod \"798531b9d9a078af26f5f153dd8093f0980ac32bb052a41050c010ef74b88wp\" (UID: \"33b3d1b9-c7e6-44e3-9259-1251e4e340a6\") " pod="openstack-operators/798531b9d9a078af26f5f153dd8093f0980ac32bb052a41050c010ef74b88wp" Dec 06 03:48:31 crc kubenswrapper[4980]: I1206 03:48:31.117600 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/33b3d1b9-c7e6-44e3-9259-1251e4e340a6-util\") pod \"798531b9d9a078af26f5f153dd8093f0980ac32bb052a41050c010ef74b88wp\" (UID: \"33b3d1b9-c7e6-44e3-9259-1251e4e340a6\") " pod="openstack-operators/798531b9d9a078af26f5f153dd8093f0980ac32bb052a41050c010ef74b88wp" Dec 06 03:48:31 crc kubenswrapper[4980]: I1206 03:48:31.117675 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k8thz\" (UniqueName: \"kubernetes.io/projected/33b3d1b9-c7e6-44e3-9259-1251e4e340a6-kube-api-access-k8thz\") pod \"798531b9d9a078af26f5f153dd8093f0980ac32bb052a41050c010ef74b88wp\" (UID: \"33b3d1b9-c7e6-44e3-9259-1251e4e340a6\") " pod="openstack-operators/798531b9d9a078af26f5f153dd8093f0980ac32bb052a41050c010ef74b88wp" Dec 06 03:48:31 crc kubenswrapper[4980]: I1206 03:48:31.117957 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/33b3d1b9-c7e6-44e3-9259-1251e4e340a6-bundle\") pod \"798531b9d9a078af26f5f153dd8093f0980ac32bb052a41050c010ef74b88wp\" (UID: \"33b3d1b9-c7e6-44e3-9259-1251e4e340a6\") " pod="openstack-operators/798531b9d9a078af26f5f153dd8093f0980ac32bb052a41050c010ef74b88wp" Dec 06 03:48:31 crc kubenswrapper[4980]: I1206 03:48:31.118059 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/33b3d1b9-c7e6-44e3-9259-1251e4e340a6-util\") pod \"798531b9d9a078af26f5f153dd8093f0980ac32bb052a41050c010ef74b88wp\" (UID: \"33b3d1b9-c7e6-44e3-9259-1251e4e340a6\") " pod="openstack-operators/798531b9d9a078af26f5f153dd8093f0980ac32bb052a41050c010ef74b88wp" Dec 06 03:48:31 crc kubenswrapper[4980]: I1206 03:48:31.145771 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k8thz\" (UniqueName: \"kubernetes.io/projected/33b3d1b9-c7e6-44e3-9259-1251e4e340a6-kube-api-access-k8thz\") pod \"798531b9d9a078af26f5f153dd8093f0980ac32bb052a41050c010ef74b88wp\" (UID: \"33b3d1b9-c7e6-44e3-9259-1251e4e340a6\") " pod="openstack-operators/798531b9d9a078af26f5f153dd8093f0980ac32bb052a41050c010ef74b88wp" Dec 06 03:48:31 crc kubenswrapper[4980]: I1206 03:48:31.198378 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/798531b9d9a078af26f5f153dd8093f0980ac32bb052a41050c010ef74b88wp" Dec 06 03:48:31 crc kubenswrapper[4980]: I1206 03:48:31.598030 4980 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/798531b9d9a078af26f5f153dd8093f0980ac32bb052a41050c010ef74b88wp"] Dec 06 03:48:31 crc kubenswrapper[4980]: W1206 03:48:31.602391 4980 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod33b3d1b9_c7e6_44e3_9259_1251e4e340a6.slice/crio-7f946f5904b81eb4c35fd5969d2e3d76c0a29559f12ca61dd0da57c04d6b6938 WatchSource:0}: Error finding container 7f946f5904b81eb4c35fd5969d2e3d76c0a29559f12ca61dd0da57c04d6b6938: Status 404 returned error can't find the container with id 7f946f5904b81eb4c35fd5969d2e3d76c0a29559f12ca61dd0da57c04d6b6938 Dec 06 03:48:31 crc kubenswrapper[4980]: I1206 03:48:31.980599 4980 generic.go:334] "Generic (PLEG): container finished" podID="33b3d1b9-c7e6-44e3-9259-1251e4e340a6" containerID="b2d3d1717e499bbf1351742c914bed01d8d8c59426274c8f418fce853909c0c4" exitCode=0 Dec 06 03:48:31 crc kubenswrapper[4980]: I1206 03:48:31.980654 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/798531b9d9a078af26f5f153dd8093f0980ac32bb052a41050c010ef74b88wp" event={"ID":"33b3d1b9-c7e6-44e3-9259-1251e4e340a6","Type":"ContainerDied","Data":"b2d3d1717e499bbf1351742c914bed01d8d8c59426274c8f418fce853909c0c4"} Dec 06 03:48:31 crc kubenswrapper[4980]: I1206 03:48:31.980684 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/798531b9d9a078af26f5f153dd8093f0980ac32bb052a41050c010ef74b88wp" event={"ID":"33b3d1b9-c7e6-44e3-9259-1251e4e340a6","Type":"ContainerStarted","Data":"7f946f5904b81eb4c35fd5969d2e3d76c0a29559f12ca61dd0da57c04d6b6938"} Dec 06 03:48:32 crc kubenswrapper[4980]: I1206 03:48:32.992007 4980 generic.go:334] "Generic (PLEG): container finished" podID="33b3d1b9-c7e6-44e3-9259-1251e4e340a6" containerID="22499427d536f88ac32b48e65fb46e2ece074f56a709898ca696eb5fb3c0b97d" exitCode=0 Dec 06 03:48:32 crc kubenswrapper[4980]: I1206 03:48:32.992135 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/798531b9d9a078af26f5f153dd8093f0980ac32bb052a41050c010ef74b88wp" event={"ID":"33b3d1b9-c7e6-44e3-9259-1251e4e340a6","Type":"ContainerDied","Data":"22499427d536f88ac32b48e65fb46e2ece074f56a709898ca696eb5fb3c0b97d"} Dec 06 03:48:34 crc kubenswrapper[4980]: I1206 03:48:34.003715 4980 generic.go:334] "Generic (PLEG): container finished" podID="33b3d1b9-c7e6-44e3-9259-1251e4e340a6" containerID="a48680bec4b52884db16c40bc2a0d854173d6e3b7a58aae0c2097e26e7cac868" exitCode=0 Dec 06 03:48:34 crc kubenswrapper[4980]: I1206 03:48:34.003827 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/798531b9d9a078af26f5f153dd8093f0980ac32bb052a41050c010ef74b88wp" event={"ID":"33b3d1b9-c7e6-44e3-9259-1251e4e340a6","Type":"ContainerDied","Data":"a48680bec4b52884db16c40bc2a0d854173d6e3b7a58aae0c2097e26e7cac868"} Dec 06 03:48:35 crc kubenswrapper[4980]: I1206 03:48:35.281434 4980 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/798531b9d9a078af26f5f153dd8093f0980ac32bb052a41050c010ef74b88wp" Dec 06 03:48:35 crc kubenswrapper[4980]: I1206 03:48:35.471349 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/33b3d1b9-c7e6-44e3-9259-1251e4e340a6-bundle\") pod \"33b3d1b9-c7e6-44e3-9259-1251e4e340a6\" (UID: \"33b3d1b9-c7e6-44e3-9259-1251e4e340a6\") " Dec 06 03:48:35 crc kubenswrapper[4980]: I1206 03:48:35.471449 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k8thz\" (UniqueName: \"kubernetes.io/projected/33b3d1b9-c7e6-44e3-9259-1251e4e340a6-kube-api-access-k8thz\") pod \"33b3d1b9-c7e6-44e3-9259-1251e4e340a6\" (UID: \"33b3d1b9-c7e6-44e3-9259-1251e4e340a6\") " Dec 06 03:48:35 crc kubenswrapper[4980]: I1206 03:48:35.471550 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/33b3d1b9-c7e6-44e3-9259-1251e4e340a6-util\") pod \"33b3d1b9-c7e6-44e3-9259-1251e4e340a6\" (UID: \"33b3d1b9-c7e6-44e3-9259-1251e4e340a6\") " Dec 06 03:48:35 crc kubenswrapper[4980]: I1206 03:48:35.472964 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/33b3d1b9-c7e6-44e3-9259-1251e4e340a6-bundle" (OuterVolumeSpecName: "bundle") pod "33b3d1b9-c7e6-44e3-9259-1251e4e340a6" (UID: "33b3d1b9-c7e6-44e3-9259-1251e4e340a6"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 03:48:35 crc kubenswrapper[4980]: I1206 03:48:35.479578 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/33b3d1b9-c7e6-44e3-9259-1251e4e340a6-kube-api-access-k8thz" (OuterVolumeSpecName: "kube-api-access-k8thz") pod "33b3d1b9-c7e6-44e3-9259-1251e4e340a6" (UID: "33b3d1b9-c7e6-44e3-9259-1251e4e340a6"). InnerVolumeSpecName "kube-api-access-k8thz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 03:48:35 crc kubenswrapper[4980]: I1206 03:48:35.489067 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/33b3d1b9-c7e6-44e3-9259-1251e4e340a6-util" (OuterVolumeSpecName: "util") pod "33b3d1b9-c7e6-44e3-9259-1251e4e340a6" (UID: "33b3d1b9-c7e6-44e3-9259-1251e4e340a6"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 03:48:35 crc kubenswrapper[4980]: I1206 03:48:35.573589 4980 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/33b3d1b9-c7e6-44e3-9259-1251e4e340a6-util\") on node \"crc\" DevicePath \"\"" Dec 06 03:48:35 crc kubenswrapper[4980]: I1206 03:48:35.573733 4980 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/33b3d1b9-c7e6-44e3-9259-1251e4e340a6-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 03:48:35 crc kubenswrapper[4980]: I1206 03:48:35.573766 4980 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k8thz\" (UniqueName: \"kubernetes.io/projected/33b3d1b9-c7e6-44e3-9259-1251e4e340a6-kube-api-access-k8thz\") on node \"crc\" DevicePath \"\"" Dec 06 03:48:36 crc kubenswrapper[4980]: I1206 03:48:36.020412 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/798531b9d9a078af26f5f153dd8093f0980ac32bb052a41050c010ef74b88wp" event={"ID":"33b3d1b9-c7e6-44e3-9259-1251e4e340a6","Type":"ContainerDied","Data":"7f946f5904b81eb4c35fd5969d2e3d76c0a29559f12ca61dd0da57c04d6b6938"} Dec 06 03:48:36 crc kubenswrapper[4980]: I1206 03:48:36.020452 4980 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7f946f5904b81eb4c35fd5969d2e3d76c0a29559f12ca61dd0da57c04d6b6938" Dec 06 03:48:36 crc kubenswrapper[4980]: I1206 03:48:36.020616 4980 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/798531b9d9a078af26f5f153dd8093f0980ac32bb052a41050c010ef74b88wp" Dec 06 03:48:44 crc kubenswrapper[4980]: I1206 03:48:44.145611 4980 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-f567f77d7-bg9qv"] Dec 06 03:48:44 crc kubenswrapper[4980]: E1206 03:48:44.146400 4980 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="33b3d1b9-c7e6-44e3-9259-1251e4e340a6" containerName="util" Dec 06 03:48:44 crc kubenswrapper[4980]: I1206 03:48:44.146417 4980 state_mem.go:107] "Deleted CPUSet assignment" podUID="33b3d1b9-c7e6-44e3-9259-1251e4e340a6" containerName="util" Dec 06 03:48:44 crc kubenswrapper[4980]: E1206 03:48:44.146434 4980 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="33b3d1b9-c7e6-44e3-9259-1251e4e340a6" containerName="pull" Dec 06 03:48:44 crc kubenswrapper[4980]: I1206 03:48:44.146439 4980 state_mem.go:107] "Deleted CPUSet assignment" podUID="33b3d1b9-c7e6-44e3-9259-1251e4e340a6" containerName="pull" Dec 06 03:48:44 crc kubenswrapper[4980]: E1206 03:48:44.146449 4980 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="33b3d1b9-c7e6-44e3-9259-1251e4e340a6" containerName="extract" Dec 06 03:48:44 crc kubenswrapper[4980]: I1206 03:48:44.146456 4980 state_mem.go:107] "Deleted CPUSet assignment" podUID="33b3d1b9-c7e6-44e3-9259-1251e4e340a6" containerName="extract" Dec 06 03:48:44 crc kubenswrapper[4980]: I1206 03:48:44.146587 4980 memory_manager.go:354] "RemoveStaleState removing state" podUID="33b3d1b9-c7e6-44e3-9259-1251e4e340a6" containerName="extract" Dec 06 03:48:44 crc kubenswrapper[4980]: I1206 03:48:44.147252 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-f567f77d7-bg9qv" Dec 06 03:48:44 crc kubenswrapper[4980]: I1206 03:48:44.150998 4980 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"mariadb-operator-controller-manager-service-cert" Dec 06 03:48:44 crc kubenswrapper[4980]: I1206 03:48:44.151093 4980 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"mariadb-operator-controller-manager-dockercfg-pq6m2" Dec 06 03:48:44 crc kubenswrapper[4980]: I1206 03:48:44.151230 4980 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"webhook-server-cert" Dec 06 03:48:44 crc kubenswrapper[4980]: I1206 03:48:44.168213 4980 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-f567f77d7-bg9qv"] Dec 06 03:48:44 crc kubenswrapper[4980]: I1206 03:48:44.224716 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/710bbcb6-d1c3-46a3-9f04-82ca8e5af2fc-webhook-cert\") pod \"mariadb-operator-controller-manager-f567f77d7-bg9qv\" (UID: \"710bbcb6-d1c3-46a3-9f04-82ca8e5af2fc\") " pod="openstack-operators/mariadb-operator-controller-manager-f567f77d7-bg9qv" Dec 06 03:48:44 crc kubenswrapper[4980]: I1206 03:48:44.224832 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4fwr7\" (UniqueName: \"kubernetes.io/projected/710bbcb6-d1c3-46a3-9f04-82ca8e5af2fc-kube-api-access-4fwr7\") pod \"mariadb-operator-controller-manager-f567f77d7-bg9qv\" (UID: \"710bbcb6-d1c3-46a3-9f04-82ca8e5af2fc\") " pod="openstack-operators/mariadb-operator-controller-manager-f567f77d7-bg9qv" Dec 06 03:48:44 crc kubenswrapper[4980]: I1206 03:48:44.224862 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/710bbcb6-d1c3-46a3-9f04-82ca8e5af2fc-apiservice-cert\") pod \"mariadb-operator-controller-manager-f567f77d7-bg9qv\" (UID: \"710bbcb6-d1c3-46a3-9f04-82ca8e5af2fc\") " pod="openstack-operators/mariadb-operator-controller-manager-f567f77d7-bg9qv" Dec 06 03:48:44 crc kubenswrapper[4980]: I1206 03:48:44.326119 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/710bbcb6-d1c3-46a3-9f04-82ca8e5af2fc-webhook-cert\") pod \"mariadb-operator-controller-manager-f567f77d7-bg9qv\" (UID: \"710bbcb6-d1c3-46a3-9f04-82ca8e5af2fc\") " pod="openstack-operators/mariadb-operator-controller-manager-f567f77d7-bg9qv" Dec 06 03:48:44 crc kubenswrapper[4980]: I1206 03:48:44.326199 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4fwr7\" (UniqueName: \"kubernetes.io/projected/710bbcb6-d1c3-46a3-9f04-82ca8e5af2fc-kube-api-access-4fwr7\") pod \"mariadb-operator-controller-manager-f567f77d7-bg9qv\" (UID: \"710bbcb6-d1c3-46a3-9f04-82ca8e5af2fc\") " pod="openstack-operators/mariadb-operator-controller-manager-f567f77d7-bg9qv" Dec 06 03:48:44 crc kubenswrapper[4980]: I1206 03:48:44.326226 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/710bbcb6-d1c3-46a3-9f04-82ca8e5af2fc-apiservice-cert\") pod \"mariadb-operator-controller-manager-f567f77d7-bg9qv\" (UID: \"710bbcb6-d1c3-46a3-9f04-82ca8e5af2fc\") " pod="openstack-operators/mariadb-operator-controller-manager-f567f77d7-bg9qv" Dec 06 03:48:44 crc kubenswrapper[4980]: I1206 03:48:44.331604 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/710bbcb6-d1c3-46a3-9f04-82ca8e5af2fc-apiservice-cert\") pod \"mariadb-operator-controller-manager-f567f77d7-bg9qv\" (UID: \"710bbcb6-d1c3-46a3-9f04-82ca8e5af2fc\") " pod="openstack-operators/mariadb-operator-controller-manager-f567f77d7-bg9qv" Dec 06 03:48:44 crc kubenswrapper[4980]: I1206 03:48:44.331649 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/710bbcb6-d1c3-46a3-9f04-82ca8e5af2fc-webhook-cert\") pod \"mariadb-operator-controller-manager-f567f77d7-bg9qv\" (UID: \"710bbcb6-d1c3-46a3-9f04-82ca8e5af2fc\") " pod="openstack-operators/mariadb-operator-controller-manager-f567f77d7-bg9qv" Dec 06 03:48:44 crc kubenswrapper[4980]: I1206 03:48:44.348893 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4fwr7\" (UniqueName: \"kubernetes.io/projected/710bbcb6-d1c3-46a3-9f04-82ca8e5af2fc-kube-api-access-4fwr7\") pod \"mariadb-operator-controller-manager-f567f77d7-bg9qv\" (UID: \"710bbcb6-d1c3-46a3-9f04-82ca8e5af2fc\") " pod="openstack-operators/mariadb-operator-controller-manager-f567f77d7-bg9qv" Dec 06 03:48:44 crc kubenswrapper[4980]: I1206 03:48:44.466052 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-f567f77d7-bg9qv" Dec 06 03:48:45 crc kubenswrapper[4980]: I1206 03:48:45.210685 4980 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-f567f77d7-bg9qv"] Dec 06 03:48:46 crc kubenswrapper[4980]: I1206 03:48:46.091239 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-f567f77d7-bg9qv" event={"ID":"710bbcb6-d1c3-46a3-9f04-82ca8e5af2fc","Type":"ContainerStarted","Data":"aa6322c20213799996b8193b66d79474fea664b554f4891fa62b4f163c7a4ccf"} Dec 06 03:48:51 crc kubenswrapper[4980]: I1206 03:48:51.136997 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-f567f77d7-bg9qv" event={"ID":"710bbcb6-d1c3-46a3-9f04-82ca8e5af2fc","Type":"ContainerStarted","Data":"38989989a7036cc4c5454c896fbb4e5c75b52abd992885dea016baa07c20a6f8"} Dec 06 03:48:53 crc kubenswrapper[4980]: I1206 03:48:53.151232 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-f567f77d7-bg9qv" event={"ID":"710bbcb6-d1c3-46a3-9f04-82ca8e5af2fc","Type":"ContainerStarted","Data":"a620a0fea6b024ac991726743f9bc4c8212da49dcb49c317d920d2178e1fa16f"} Dec 06 03:48:53 crc kubenswrapper[4980]: I1206 03:48:53.151563 4980 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/mariadb-operator-controller-manager-f567f77d7-bg9qv" Dec 06 03:48:53 crc kubenswrapper[4980]: I1206 03:48:53.186893 4980 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/mariadb-operator-controller-manager-f567f77d7-bg9qv" podStartSLOduration=1.902270707 podStartE2EDuration="9.186859338s" podCreationTimestamp="2025-12-06 03:48:44 +0000 UTC" firstStartedPulling="2025-12-06 03:48:45.218101422 +0000 UTC m=+884.462852693" lastFinishedPulling="2025-12-06 03:48:52.502690053 +0000 UTC m=+891.747441324" observedRunningTime="2025-12-06 03:48:53.183372868 +0000 UTC m=+892.428124149" watchObservedRunningTime="2025-12-06 03:48:53.186859338 +0000 UTC m=+892.431610609" Dec 06 03:49:04 crc kubenswrapper[4980]: I1206 03:49:04.471865 4980 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/mariadb-operator-controller-manager-f567f77d7-bg9qv" Dec 06 03:49:05 crc kubenswrapper[4980]: I1206 03:49:05.040241 4980 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/infra-operator-index-jj8dz"] Dec 06 03:49:05 crc kubenswrapper[4980]: I1206 03:49:05.041172 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-index-jj8dz" Dec 06 03:49:05 crc kubenswrapper[4980]: I1206 03:49:05.046317 4980 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-index-dockercfg-87f4h" Dec 06 03:49:05 crc kubenswrapper[4980]: I1206 03:49:05.122353 4980 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-index-jj8dz"] Dec 06 03:49:05 crc kubenswrapper[4980]: I1206 03:49:05.160825 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xk64t\" (UniqueName: \"kubernetes.io/projected/998e7f3b-ecfb-47dd-8d28-f55cf9e75526-kube-api-access-xk64t\") pod \"infra-operator-index-jj8dz\" (UID: \"998e7f3b-ecfb-47dd-8d28-f55cf9e75526\") " pod="openstack-operators/infra-operator-index-jj8dz" Dec 06 03:49:05 crc kubenswrapper[4980]: I1206 03:49:05.262524 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xk64t\" (UniqueName: \"kubernetes.io/projected/998e7f3b-ecfb-47dd-8d28-f55cf9e75526-kube-api-access-xk64t\") pod \"infra-operator-index-jj8dz\" (UID: \"998e7f3b-ecfb-47dd-8d28-f55cf9e75526\") " pod="openstack-operators/infra-operator-index-jj8dz" Dec 06 03:49:05 crc kubenswrapper[4980]: I1206 03:49:05.292294 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xk64t\" (UniqueName: \"kubernetes.io/projected/998e7f3b-ecfb-47dd-8d28-f55cf9e75526-kube-api-access-xk64t\") pod \"infra-operator-index-jj8dz\" (UID: \"998e7f3b-ecfb-47dd-8d28-f55cf9e75526\") " pod="openstack-operators/infra-operator-index-jj8dz" Dec 06 03:49:05 crc kubenswrapper[4980]: I1206 03:49:05.357414 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-index-jj8dz" Dec 06 03:49:05 crc kubenswrapper[4980]: I1206 03:49:05.886299 4980 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-index-jj8dz"] Dec 06 03:49:06 crc kubenswrapper[4980]: I1206 03:49:06.228021 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-index-jj8dz" event={"ID":"998e7f3b-ecfb-47dd-8d28-f55cf9e75526","Type":"ContainerStarted","Data":"eb2b5ea434634a294b5ca0f86a11e54d075be3bab7bf907a99e615f2396b8b20"} Dec 06 03:49:07 crc kubenswrapper[4980]: I1206 03:49:07.820543 4980 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/infra-operator-index-jj8dz"] Dec 06 03:49:08 crc kubenswrapper[4980]: I1206 03:49:08.226491 4980 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/infra-operator-index-gwm9t"] Dec 06 03:49:08 crc kubenswrapper[4980]: I1206 03:49:08.227611 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-index-gwm9t" Dec 06 03:49:08 crc kubenswrapper[4980]: I1206 03:49:08.233940 4980 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-index-gwm9t"] Dec 06 03:49:08 crc kubenswrapper[4980]: I1206 03:49:08.387244 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q64f6\" (UniqueName: \"kubernetes.io/projected/8999e732-bb13-4416-b63e-c54355c9f489-kube-api-access-q64f6\") pod \"infra-operator-index-gwm9t\" (UID: \"8999e732-bb13-4416-b63e-c54355c9f489\") " pod="openstack-operators/infra-operator-index-gwm9t" Dec 06 03:49:08 crc kubenswrapper[4980]: I1206 03:49:08.488955 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q64f6\" (UniqueName: \"kubernetes.io/projected/8999e732-bb13-4416-b63e-c54355c9f489-kube-api-access-q64f6\") pod \"infra-operator-index-gwm9t\" (UID: \"8999e732-bb13-4416-b63e-c54355c9f489\") " pod="openstack-operators/infra-operator-index-gwm9t" Dec 06 03:49:08 crc kubenswrapper[4980]: I1206 03:49:08.507855 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q64f6\" (UniqueName: \"kubernetes.io/projected/8999e732-bb13-4416-b63e-c54355c9f489-kube-api-access-q64f6\") pod \"infra-operator-index-gwm9t\" (UID: \"8999e732-bb13-4416-b63e-c54355c9f489\") " pod="openstack-operators/infra-operator-index-gwm9t" Dec 06 03:49:08 crc kubenswrapper[4980]: I1206 03:49:08.551810 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-index-gwm9t" Dec 06 03:49:09 crc kubenswrapper[4980]: I1206 03:49:09.131165 4980 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-index-gwm9t"] Dec 06 03:49:11 crc kubenswrapper[4980]: W1206 03:49:11.726100 4980 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8999e732_bb13_4416_b63e_c54355c9f489.slice/crio-e4db23ad82af39bcd719d83dd66dd7a50e0eddd813ae0c5e1fd41573c911c3c1 WatchSource:0}: Error finding container e4db23ad82af39bcd719d83dd66dd7a50e0eddd813ae0c5e1fd41573c911c3c1: Status 404 returned error can't find the container with id e4db23ad82af39bcd719d83dd66dd7a50e0eddd813ae0c5e1fd41573c911c3c1 Dec 06 03:49:12 crc kubenswrapper[4980]: I1206 03:49:12.319777 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-index-gwm9t" event={"ID":"8999e732-bb13-4416-b63e-c54355c9f489","Type":"ContainerStarted","Data":"e4db23ad82af39bcd719d83dd66dd7a50e0eddd813ae0c5e1fd41573c911c3c1"} Dec 06 03:49:19 crc kubenswrapper[4980]: I1206 03:49:19.376318 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-index-jj8dz" event={"ID":"998e7f3b-ecfb-47dd-8d28-f55cf9e75526","Type":"ContainerStarted","Data":"9952aa79997f8a03b08facbec3347bddf7f633d03037fc10fc1900e4657184f2"} Dec 06 03:49:19 crc kubenswrapper[4980]: I1206 03:49:19.376484 4980 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/infra-operator-index-jj8dz" podUID="998e7f3b-ecfb-47dd-8d28-f55cf9e75526" containerName="registry-server" containerID="cri-o://9952aa79997f8a03b08facbec3347bddf7f633d03037fc10fc1900e4657184f2" gracePeriod=2 Dec 06 03:49:19 crc kubenswrapper[4980]: I1206 03:49:19.378160 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-index-gwm9t" event={"ID":"8999e732-bb13-4416-b63e-c54355c9f489","Type":"ContainerStarted","Data":"48e267668a919f3b4a5e300c5a05482c9b8dd6550bf2902a1114bcab93f14c0f"} Dec 06 03:49:19 crc kubenswrapper[4980]: I1206 03:49:19.398418 4980 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/infra-operator-index-jj8dz" podStartSLOduration=1.641666812 podStartE2EDuration="14.398401404s" podCreationTimestamp="2025-12-06 03:49:05 +0000 UTC" firstStartedPulling="2025-12-06 03:49:05.902586771 +0000 UTC m=+905.147338042" lastFinishedPulling="2025-12-06 03:49:18.659321363 +0000 UTC m=+917.904072634" observedRunningTime="2025-12-06 03:49:19.39408316 +0000 UTC m=+918.638834431" watchObservedRunningTime="2025-12-06 03:49:19.398401404 +0000 UTC m=+918.643152675" Dec 06 03:49:19 crc kubenswrapper[4980]: I1206 03:49:19.413476 4980 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/infra-operator-index-gwm9t" podStartSLOduration=4.565385523 podStartE2EDuration="11.413459298s" podCreationTimestamp="2025-12-06 03:49:08 +0000 UTC" firstStartedPulling="2025-12-06 03:49:11.729070931 +0000 UTC m=+910.973822212" lastFinishedPulling="2025-12-06 03:49:18.577144676 +0000 UTC m=+917.821895987" observedRunningTime="2025-12-06 03:49:19.408543616 +0000 UTC m=+918.653294887" watchObservedRunningTime="2025-12-06 03:49:19.413459298 +0000 UTC m=+918.658210569" Dec 06 03:49:19 crc kubenswrapper[4980]: I1206 03:49:19.814699 4980 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-index-jj8dz" Dec 06 03:49:19 crc kubenswrapper[4980]: I1206 03:49:19.918609 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xk64t\" (UniqueName: \"kubernetes.io/projected/998e7f3b-ecfb-47dd-8d28-f55cf9e75526-kube-api-access-xk64t\") pod \"998e7f3b-ecfb-47dd-8d28-f55cf9e75526\" (UID: \"998e7f3b-ecfb-47dd-8d28-f55cf9e75526\") " Dec 06 03:49:19 crc kubenswrapper[4980]: I1206 03:49:19.928763 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/998e7f3b-ecfb-47dd-8d28-f55cf9e75526-kube-api-access-xk64t" (OuterVolumeSpecName: "kube-api-access-xk64t") pod "998e7f3b-ecfb-47dd-8d28-f55cf9e75526" (UID: "998e7f3b-ecfb-47dd-8d28-f55cf9e75526"). InnerVolumeSpecName "kube-api-access-xk64t". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 03:49:20 crc kubenswrapper[4980]: I1206 03:49:20.021310 4980 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xk64t\" (UniqueName: \"kubernetes.io/projected/998e7f3b-ecfb-47dd-8d28-f55cf9e75526-kube-api-access-xk64t\") on node \"crc\" DevicePath \"\"" Dec 06 03:49:20 crc kubenswrapper[4980]: I1206 03:49:20.386695 4980 generic.go:334] "Generic (PLEG): container finished" podID="998e7f3b-ecfb-47dd-8d28-f55cf9e75526" containerID="9952aa79997f8a03b08facbec3347bddf7f633d03037fc10fc1900e4657184f2" exitCode=0 Dec 06 03:49:20 crc kubenswrapper[4980]: I1206 03:49:20.386773 4980 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-index-jj8dz" Dec 06 03:49:20 crc kubenswrapper[4980]: I1206 03:49:20.386805 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-index-jj8dz" event={"ID":"998e7f3b-ecfb-47dd-8d28-f55cf9e75526","Type":"ContainerDied","Data":"9952aa79997f8a03b08facbec3347bddf7f633d03037fc10fc1900e4657184f2"} Dec 06 03:49:20 crc kubenswrapper[4980]: I1206 03:49:20.386871 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-index-jj8dz" event={"ID":"998e7f3b-ecfb-47dd-8d28-f55cf9e75526","Type":"ContainerDied","Data":"eb2b5ea434634a294b5ca0f86a11e54d075be3bab7bf907a99e615f2396b8b20"} Dec 06 03:49:20 crc kubenswrapper[4980]: I1206 03:49:20.386950 4980 scope.go:117] "RemoveContainer" containerID="9952aa79997f8a03b08facbec3347bddf7f633d03037fc10fc1900e4657184f2" Dec 06 03:49:20 crc kubenswrapper[4980]: I1206 03:49:20.402128 4980 scope.go:117] "RemoveContainer" containerID="9952aa79997f8a03b08facbec3347bddf7f633d03037fc10fc1900e4657184f2" Dec 06 03:49:20 crc kubenswrapper[4980]: E1206 03:49:20.402863 4980 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9952aa79997f8a03b08facbec3347bddf7f633d03037fc10fc1900e4657184f2\": container with ID starting with 9952aa79997f8a03b08facbec3347bddf7f633d03037fc10fc1900e4657184f2 not found: ID does not exist" containerID="9952aa79997f8a03b08facbec3347bddf7f633d03037fc10fc1900e4657184f2" Dec 06 03:49:20 crc kubenswrapper[4980]: I1206 03:49:20.402928 4980 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9952aa79997f8a03b08facbec3347bddf7f633d03037fc10fc1900e4657184f2"} err="failed to get container status \"9952aa79997f8a03b08facbec3347bddf7f633d03037fc10fc1900e4657184f2\": rpc error: code = NotFound desc = could not find container \"9952aa79997f8a03b08facbec3347bddf7f633d03037fc10fc1900e4657184f2\": container with ID starting with 9952aa79997f8a03b08facbec3347bddf7f633d03037fc10fc1900e4657184f2 not found: ID does not exist" Dec 06 03:49:20 crc kubenswrapper[4980]: I1206 03:49:20.419609 4980 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/infra-operator-index-jj8dz"] Dec 06 03:49:20 crc kubenswrapper[4980]: I1206 03:49:20.423371 4980 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/infra-operator-index-jj8dz"] Dec 06 03:49:21 crc kubenswrapper[4980]: I1206 03:49:21.191652 4980 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="998e7f3b-ecfb-47dd-8d28-f55cf9e75526" path="/var/lib/kubelet/pods/998e7f3b-ecfb-47dd-8d28-f55cf9e75526/volumes" Dec 06 03:49:25 crc kubenswrapper[4980]: I1206 03:49:25.137762 4980 patch_prober.go:28] interesting pod/machine-config-daemon-r5zfb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 03:49:25 crc kubenswrapper[4980]: I1206 03:49:25.138177 4980 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-r5zfb" podUID="320f44d1-a671-4a91-b328-a8b0fdd8f23a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 03:49:28 crc kubenswrapper[4980]: I1206 03:49:28.552440 4980 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/infra-operator-index-gwm9t" Dec 06 03:49:28 crc kubenswrapper[4980]: I1206 03:49:28.553724 4980 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/infra-operator-index-gwm9t" Dec 06 03:49:28 crc kubenswrapper[4980]: I1206 03:49:28.596656 4980 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/infra-operator-index-gwm9t" Dec 06 03:49:29 crc kubenswrapper[4980]: I1206 03:49:29.547945 4980 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/infra-operator-index-gwm9t" Dec 06 03:49:36 crc kubenswrapper[4980]: I1206 03:49:36.273330 4980 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/8ad853bb090b3a36d4332bde14850afdc9872f13e5d95ff50094430eb62kjwk"] Dec 06 03:49:36 crc kubenswrapper[4980]: E1206 03:49:36.275346 4980 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="998e7f3b-ecfb-47dd-8d28-f55cf9e75526" containerName="registry-server" Dec 06 03:49:36 crc kubenswrapper[4980]: I1206 03:49:36.275434 4980 state_mem.go:107] "Deleted CPUSet assignment" podUID="998e7f3b-ecfb-47dd-8d28-f55cf9e75526" containerName="registry-server" Dec 06 03:49:36 crc kubenswrapper[4980]: I1206 03:49:36.275899 4980 memory_manager.go:354] "RemoveStaleState removing state" podUID="998e7f3b-ecfb-47dd-8d28-f55cf9e75526" containerName="registry-server" Dec 06 03:49:36 crc kubenswrapper[4980]: I1206 03:49:36.277027 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/8ad853bb090b3a36d4332bde14850afdc9872f13e5d95ff50094430eb62kjwk" Dec 06 03:49:36 crc kubenswrapper[4980]: I1206 03:49:36.279582 4980 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"default-dockercfg-brs8h" Dec 06 03:49:36 crc kubenswrapper[4980]: I1206 03:49:36.293929 4980 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/8ad853bb090b3a36d4332bde14850afdc9872f13e5d95ff50094430eb62kjwk"] Dec 06 03:49:36 crc kubenswrapper[4980]: I1206 03:49:36.335603 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/a8036768-6216-4f7a-adc6-7f81287574e3-util\") pod \"8ad853bb090b3a36d4332bde14850afdc9872f13e5d95ff50094430eb62kjwk\" (UID: \"a8036768-6216-4f7a-adc6-7f81287574e3\") " pod="openstack-operators/8ad853bb090b3a36d4332bde14850afdc9872f13e5d95ff50094430eb62kjwk" Dec 06 03:49:36 crc kubenswrapper[4980]: I1206 03:49:36.335705 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q52mg\" (UniqueName: \"kubernetes.io/projected/a8036768-6216-4f7a-adc6-7f81287574e3-kube-api-access-q52mg\") pod \"8ad853bb090b3a36d4332bde14850afdc9872f13e5d95ff50094430eb62kjwk\" (UID: \"a8036768-6216-4f7a-adc6-7f81287574e3\") " pod="openstack-operators/8ad853bb090b3a36d4332bde14850afdc9872f13e5d95ff50094430eb62kjwk" Dec 06 03:49:36 crc kubenswrapper[4980]: I1206 03:49:36.335784 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/a8036768-6216-4f7a-adc6-7f81287574e3-bundle\") pod \"8ad853bb090b3a36d4332bde14850afdc9872f13e5d95ff50094430eb62kjwk\" (UID: \"a8036768-6216-4f7a-adc6-7f81287574e3\") " pod="openstack-operators/8ad853bb090b3a36d4332bde14850afdc9872f13e5d95ff50094430eb62kjwk" Dec 06 03:49:36 crc kubenswrapper[4980]: I1206 03:49:36.436574 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/a8036768-6216-4f7a-adc6-7f81287574e3-util\") pod \"8ad853bb090b3a36d4332bde14850afdc9872f13e5d95ff50094430eb62kjwk\" (UID: \"a8036768-6216-4f7a-adc6-7f81287574e3\") " pod="openstack-operators/8ad853bb090b3a36d4332bde14850afdc9872f13e5d95ff50094430eb62kjwk" Dec 06 03:49:36 crc kubenswrapper[4980]: I1206 03:49:36.436782 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q52mg\" (UniqueName: \"kubernetes.io/projected/a8036768-6216-4f7a-adc6-7f81287574e3-kube-api-access-q52mg\") pod \"8ad853bb090b3a36d4332bde14850afdc9872f13e5d95ff50094430eb62kjwk\" (UID: \"a8036768-6216-4f7a-adc6-7f81287574e3\") " pod="openstack-operators/8ad853bb090b3a36d4332bde14850afdc9872f13e5d95ff50094430eb62kjwk" Dec 06 03:49:36 crc kubenswrapper[4980]: I1206 03:49:36.436837 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/a8036768-6216-4f7a-adc6-7f81287574e3-bundle\") pod \"8ad853bb090b3a36d4332bde14850afdc9872f13e5d95ff50094430eb62kjwk\" (UID: \"a8036768-6216-4f7a-adc6-7f81287574e3\") " pod="openstack-operators/8ad853bb090b3a36d4332bde14850afdc9872f13e5d95ff50094430eb62kjwk" Dec 06 03:49:36 crc kubenswrapper[4980]: I1206 03:49:36.437459 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/a8036768-6216-4f7a-adc6-7f81287574e3-bundle\") pod \"8ad853bb090b3a36d4332bde14850afdc9872f13e5d95ff50094430eb62kjwk\" (UID: \"a8036768-6216-4f7a-adc6-7f81287574e3\") " pod="openstack-operators/8ad853bb090b3a36d4332bde14850afdc9872f13e5d95ff50094430eb62kjwk" Dec 06 03:49:36 crc kubenswrapper[4980]: I1206 03:49:36.437476 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/a8036768-6216-4f7a-adc6-7f81287574e3-util\") pod \"8ad853bb090b3a36d4332bde14850afdc9872f13e5d95ff50094430eb62kjwk\" (UID: \"a8036768-6216-4f7a-adc6-7f81287574e3\") " pod="openstack-operators/8ad853bb090b3a36d4332bde14850afdc9872f13e5d95ff50094430eb62kjwk" Dec 06 03:49:36 crc kubenswrapper[4980]: I1206 03:49:36.453742 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q52mg\" (UniqueName: \"kubernetes.io/projected/a8036768-6216-4f7a-adc6-7f81287574e3-kube-api-access-q52mg\") pod \"8ad853bb090b3a36d4332bde14850afdc9872f13e5d95ff50094430eb62kjwk\" (UID: \"a8036768-6216-4f7a-adc6-7f81287574e3\") " pod="openstack-operators/8ad853bb090b3a36d4332bde14850afdc9872f13e5d95ff50094430eb62kjwk" Dec 06 03:49:36 crc kubenswrapper[4980]: I1206 03:49:36.637868 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/8ad853bb090b3a36d4332bde14850afdc9872f13e5d95ff50094430eb62kjwk" Dec 06 03:49:36 crc kubenswrapper[4980]: I1206 03:49:36.860263 4980 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/8ad853bb090b3a36d4332bde14850afdc9872f13e5d95ff50094430eb62kjwk"] Dec 06 03:49:37 crc kubenswrapper[4980]: I1206 03:49:37.580708 4980 generic.go:334] "Generic (PLEG): container finished" podID="a8036768-6216-4f7a-adc6-7f81287574e3" containerID="445ba35f204990fcf971e6508ac3dd75127a715deb6848138d2b989c4f032924" exitCode=0 Dec 06 03:49:37 crc kubenswrapper[4980]: I1206 03:49:37.580838 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/8ad853bb090b3a36d4332bde14850afdc9872f13e5d95ff50094430eb62kjwk" event={"ID":"a8036768-6216-4f7a-adc6-7f81287574e3","Type":"ContainerDied","Data":"445ba35f204990fcf971e6508ac3dd75127a715deb6848138d2b989c4f032924"} Dec 06 03:49:37 crc kubenswrapper[4980]: I1206 03:49:37.581284 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/8ad853bb090b3a36d4332bde14850afdc9872f13e5d95ff50094430eb62kjwk" event={"ID":"a8036768-6216-4f7a-adc6-7f81287574e3","Type":"ContainerStarted","Data":"0f60cd1d89124663dfacf7aa12a30bea44d24c0b15eb15bd66cf9cb46640671f"} Dec 06 03:49:38 crc kubenswrapper[4980]: I1206 03:49:38.590998 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/8ad853bb090b3a36d4332bde14850afdc9872f13e5d95ff50094430eb62kjwk" event={"ID":"a8036768-6216-4f7a-adc6-7f81287574e3","Type":"ContainerStarted","Data":"bf9641b368c2306fbfbcc4cbb4bea53fc6ae355d508a0e713cc14873c9c5b3cd"} Dec 06 03:49:39 crc kubenswrapper[4980]: I1206 03:49:39.625543 4980 generic.go:334] "Generic (PLEG): container finished" podID="a8036768-6216-4f7a-adc6-7f81287574e3" containerID="bf9641b368c2306fbfbcc4cbb4bea53fc6ae355d508a0e713cc14873c9c5b3cd" exitCode=0 Dec 06 03:49:39 crc kubenswrapper[4980]: I1206 03:49:39.625594 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/8ad853bb090b3a36d4332bde14850afdc9872f13e5d95ff50094430eb62kjwk" event={"ID":"a8036768-6216-4f7a-adc6-7f81287574e3","Type":"ContainerDied","Data":"bf9641b368c2306fbfbcc4cbb4bea53fc6ae355d508a0e713cc14873c9c5b3cd"} Dec 06 03:49:40 crc kubenswrapper[4980]: I1206 03:49:40.636082 4980 generic.go:334] "Generic (PLEG): container finished" podID="a8036768-6216-4f7a-adc6-7f81287574e3" containerID="f9c74d3b56c6de8e7ef473703d4fb196ca0702bdd636f3a4be313af33d4acd0b" exitCode=0 Dec 06 03:49:40 crc kubenswrapper[4980]: I1206 03:49:40.636159 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/8ad853bb090b3a36d4332bde14850afdc9872f13e5d95ff50094430eb62kjwk" event={"ID":"a8036768-6216-4f7a-adc6-7f81287574e3","Type":"ContainerDied","Data":"f9c74d3b56c6de8e7ef473703d4fb196ca0702bdd636f3a4be313af33d4acd0b"} Dec 06 03:49:41 crc kubenswrapper[4980]: I1206 03:49:41.942752 4980 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/8ad853bb090b3a36d4332bde14850afdc9872f13e5d95ff50094430eb62kjwk" Dec 06 03:49:42 crc kubenswrapper[4980]: I1206 03:49:42.109260 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/a8036768-6216-4f7a-adc6-7f81287574e3-bundle\") pod \"a8036768-6216-4f7a-adc6-7f81287574e3\" (UID: \"a8036768-6216-4f7a-adc6-7f81287574e3\") " Dec 06 03:49:42 crc kubenswrapper[4980]: I1206 03:49:42.109339 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q52mg\" (UniqueName: \"kubernetes.io/projected/a8036768-6216-4f7a-adc6-7f81287574e3-kube-api-access-q52mg\") pod \"a8036768-6216-4f7a-adc6-7f81287574e3\" (UID: \"a8036768-6216-4f7a-adc6-7f81287574e3\") " Dec 06 03:49:42 crc kubenswrapper[4980]: I1206 03:49:42.109372 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/a8036768-6216-4f7a-adc6-7f81287574e3-util\") pod \"a8036768-6216-4f7a-adc6-7f81287574e3\" (UID: \"a8036768-6216-4f7a-adc6-7f81287574e3\") " Dec 06 03:49:42 crc kubenswrapper[4980]: I1206 03:49:42.110750 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a8036768-6216-4f7a-adc6-7f81287574e3-bundle" (OuterVolumeSpecName: "bundle") pod "a8036768-6216-4f7a-adc6-7f81287574e3" (UID: "a8036768-6216-4f7a-adc6-7f81287574e3"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 03:49:42 crc kubenswrapper[4980]: I1206 03:49:42.114599 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a8036768-6216-4f7a-adc6-7f81287574e3-kube-api-access-q52mg" (OuterVolumeSpecName: "kube-api-access-q52mg") pod "a8036768-6216-4f7a-adc6-7f81287574e3" (UID: "a8036768-6216-4f7a-adc6-7f81287574e3"). InnerVolumeSpecName "kube-api-access-q52mg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 03:49:42 crc kubenswrapper[4980]: I1206 03:49:42.130750 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a8036768-6216-4f7a-adc6-7f81287574e3-util" (OuterVolumeSpecName: "util") pod "a8036768-6216-4f7a-adc6-7f81287574e3" (UID: "a8036768-6216-4f7a-adc6-7f81287574e3"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 03:49:42 crc kubenswrapper[4980]: I1206 03:49:42.210805 4980 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/a8036768-6216-4f7a-adc6-7f81287574e3-util\") on node \"crc\" DevicePath \"\"" Dec 06 03:49:42 crc kubenswrapper[4980]: I1206 03:49:42.210841 4980 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/a8036768-6216-4f7a-adc6-7f81287574e3-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 03:49:42 crc kubenswrapper[4980]: I1206 03:49:42.210853 4980 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q52mg\" (UniqueName: \"kubernetes.io/projected/a8036768-6216-4f7a-adc6-7f81287574e3-kube-api-access-q52mg\") on node \"crc\" DevicePath \"\"" Dec 06 03:49:42 crc kubenswrapper[4980]: I1206 03:49:42.656310 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/8ad853bb090b3a36d4332bde14850afdc9872f13e5d95ff50094430eb62kjwk" event={"ID":"a8036768-6216-4f7a-adc6-7f81287574e3","Type":"ContainerDied","Data":"0f60cd1d89124663dfacf7aa12a30bea44d24c0b15eb15bd66cf9cb46640671f"} Dec 06 03:49:42 crc kubenswrapper[4980]: I1206 03:49:42.656368 4980 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0f60cd1d89124663dfacf7aa12a30bea44d24c0b15eb15bd66cf9cb46640671f" Dec 06 03:49:42 crc kubenswrapper[4980]: I1206 03:49:42.656481 4980 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/8ad853bb090b3a36d4332bde14850afdc9872f13e5d95ff50094430eb62kjwk" Dec 06 03:49:50 crc kubenswrapper[4980]: I1206 03:49:50.050993 4980 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/infra-operator-controller-manager-99587f5d7-rcrhn"] Dec 06 03:49:50 crc kubenswrapper[4980]: E1206 03:49:50.051815 4980 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a8036768-6216-4f7a-adc6-7f81287574e3" containerName="pull" Dec 06 03:49:50 crc kubenswrapper[4980]: I1206 03:49:50.051833 4980 state_mem.go:107] "Deleted CPUSet assignment" podUID="a8036768-6216-4f7a-adc6-7f81287574e3" containerName="pull" Dec 06 03:49:50 crc kubenswrapper[4980]: E1206 03:49:50.051861 4980 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a8036768-6216-4f7a-adc6-7f81287574e3" containerName="extract" Dec 06 03:49:50 crc kubenswrapper[4980]: I1206 03:49:50.051869 4980 state_mem.go:107] "Deleted CPUSet assignment" podUID="a8036768-6216-4f7a-adc6-7f81287574e3" containerName="extract" Dec 06 03:49:50 crc kubenswrapper[4980]: E1206 03:49:50.051883 4980 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a8036768-6216-4f7a-adc6-7f81287574e3" containerName="util" Dec 06 03:49:50 crc kubenswrapper[4980]: I1206 03:49:50.051891 4980 state_mem.go:107] "Deleted CPUSet assignment" podUID="a8036768-6216-4f7a-adc6-7f81287574e3" containerName="util" Dec 06 03:49:50 crc kubenswrapper[4980]: I1206 03:49:50.052024 4980 memory_manager.go:354] "RemoveStaleState removing state" podUID="a8036768-6216-4f7a-adc6-7f81287574e3" containerName="extract" Dec 06 03:49:50 crc kubenswrapper[4980]: I1206 03:49:50.052818 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-99587f5d7-rcrhn" Dec 06 03:49:50 crc kubenswrapper[4980]: I1206 03:49:50.054565 4980 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-controller-manager-dockercfg-qtnc2" Dec 06 03:49:50 crc kubenswrapper[4980]: I1206 03:49:50.058228 4980 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-controller-manager-service-cert" Dec 06 03:49:50 crc kubenswrapper[4980]: I1206 03:49:50.067972 4980 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-99587f5d7-rcrhn"] Dec 06 03:49:50 crc kubenswrapper[4980]: I1206 03:49:50.214644 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v6cks\" (UniqueName: \"kubernetes.io/projected/0cc75ddc-77ef-4c34-b7d1-c5e98f7666dc-kube-api-access-v6cks\") pod \"infra-operator-controller-manager-99587f5d7-rcrhn\" (UID: \"0cc75ddc-77ef-4c34-b7d1-c5e98f7666dc\") " pod="openstack-operators/infra-operator-controller-manager-99587f5d7-rcrhn" Dec 06 03:49:50 crc kubenswrapper[4980]: I1206 03:49:50.214704 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/0cc75ddc-77ef-4c34-b7d1-c5e98f7666dc-apiservice-cert\") pod \"infra-operator-controller-manager-99587f5d7-rcrhn\" (UID: \"0cc75ddc-77ef-4c34-b7d1-c5e98f7666dc\") " pod="openstack-operators/infra-operator-controller-manager-99587f5d7-rcrhn" Dec 06 03:49:50 crc kubenswrapper[4980]: I1206 03:49:50.214728 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/0cc75ddc-77ef-4c34-b7d1-c5e98f7666dc-webhook-cert\") pod \"infra-operator-controller-manager-99587f5d7-rcrhn\" (UID: \"0cc75ddc-77ef-4c34-b7d1-c5e98f7666dc\") " pod="openstack-operators/infra-operator-controller-manager-99587f5d7-rcrhn" Dec 06 03:49:50 crc kubenswrapper[4980]: I1206 03:49:50.316221 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v6cks\" (UniqueName: \"kubernetes.io/projected/0cc75ddc-77ef-4c34-b7d1-c5e98f7666dc-kube-api-access-v6cks\") pod \"infra-operator-controller-manager-99587f5d7-rcrhn\" (UID: \"0cc75ddc-77ef-4c34-b7d1-c5e98f7666dc\") " pod="openstack-operators/infra-operator-controller-manager-99587f5d7-rcrhn" Dec 06 03:49:50 crc kubenswrapper[4980]: I1206 03:49:50.316714 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/0cc75ddc-77ef-4c34-b7d1-c5e98f7666dc-apiservice-cert\") pod \"infra-operator-controller-manager-99587f5d7-rcrhn\" (UID: \"0cc75ddc-77ef-4c34-b7d1-c5e98f7666dc\") " pod="openstack-operators/infra-operator-controller-manager-99587f5d7-rcrhn" Dec 06 03:49:50 crc kubenswrapper[4980]: I1206 03:49:50.316855 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/0cc75ddc-77ef-4c34-b7d1-c5e98f7666dc-webhook-cert\") pod \"infra-operator-controller-manager-99587f5d7-rcrhn\" (UID: \"0cc75ddc-77ef-4c34-b7d1-c5e98f7666dc\") " pod="openstack-operators/infra-operator-controller-manager-99587f5d7-rcrhn" Dec 06 03:49:50 crc kubenswrapper[4980]: I1206 03:49:50.324833 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/0cc75ddc-77ef-4c34-b7d1-c5e98f7666dc-webhook-cert\") pod \"infra-operator-controller-manager-99587f5d7-rcrhn\" (UID: \"0cc75ddc-77ef-4c34-b7d1-c5e98f7666dc\") " pod="openstack-operators/infra-operator-controller-manager-99587f5d7-rcrhn" Dec 06 03:49:50 crc kubenswrapper[4980]: I1206 03:49:50.327096 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/0cc75ddc-77ef-4c34-b7d1-c5e98f7666dc-apiservice-cert\") pod \"infra-operator-controller-manager-99587f5d7-rcrhn\" (UID: \"0cc75ddc-77ef-4c34-b7d1-c5e98f7666dc\") " pod="openstack-operators/infra-operator-controller-manager-99587f5d7-rcrhn" Dec 06 03:49:50 crc kubenswrapper[4980]: I1206 03:49:50.360072 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v6cks\" (UniqueName: \"kubernetes.io/projected/0cc75ddc-77ef-4c34-b7d1-c5e98f7666dc-kube-api-access-v6cks\") pod \"infra-operator-controller-manager-99587f5d7-rcrhn\" (UID: \"0cc75ddc-77ef-4c34-b7d1-c5e98f7666dc\") " pod="openstack-operators/infra-operator-controller-manager-99587f5d7-rcrhn" Dec 06 03:49:50 crc kubenswrapper[4980]: I1206 03:49:50.368857 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-99587f5d7-rcrhn" Dec 06 03:49:50 crc kubenswrapper[4980]: I1206 03:49:50.582535 4980 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-99587f5d7-rcrhn"] Dec 06 03:49:50 crc kubenswrapper[4980]: I1206 03:49:50.699935 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-99587f5d7-rcrhn" event={"ID":"0cc75ddc-77ef-4c34-b7d1-c5e98f7666dc","Type":"ContainerStarted","Data":"9aa6949f1289f88cfbd65a45d1c61a436bdedbfc8a64fb0653b705b12b8914c6"} Dec 06 03:49:51 crc kubenswrapper[4980]: I1206 03:49:51.845587 4980 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/openstack-galera-0"] Dec 06 03:49:51 crc kubenswrapper[4980]: I1206 03:49:51.846585 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/openstack-galera-0" Dec 06 03:49:51 crc kubenswrapper[4980]: I1206 03:49:51.859398 4980 reflector.go:368] Caches populated for *v1.ConfigMap from object-"glance-kuttl-tests"/"openshift-service-ca.crt" Dec 06 03:49:51 crc kubenswrapper[4980]: I1206 03:49:51.859688 4980 reflector.go:368] Caches populated for *v1.ConfigMap from object-"glance-kuttl-tests"/"kube-root-ca.crt" Dec 06 03:49:51 crc kubenswrapper[4980]: I1206 03:49:51.859879 4980 reflector.go:368] Caches populated for *v1.ConfigMap from object-"glance-kuttl-tests"/"openstack-scripts" Dec 06 03:49:51 crc kubenswrapper[4980]: I1206 03:49:51.860133 4980 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"osp-secret" Dec 06 03:49:51 crc kubenswrapper[4980]: I1206 03:49:51.860643 4980 reflector.go:368] Caches populated for *v1.ConfigMap from object-"glance-kuttl-tests"/"openstack-config-data" Dec 06 03:49:51 crc kubenswrapper[4980]: I1206 03:49:51.860752 4980 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"galera-openstack-dockercfg-sbxwc" Dec 06 03:49:51 crc kubenswrapper[4980]: I1206 03:49:51.868991 4980 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/openstack-galera-1"] Dec 06 03:49:51 crc kubenswrapper[4980]: I1206 03:49:51.869990 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/openstack-galera-1" Dec 06 03:49:51 crc kubenswrapper[4980]: I1206 03:49:51.874284 4980 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/openstack-galera-2"] Dec 06 03:49:51 crc kubenswrapper[4980]: I1206 03:49:51.877314 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/openstack-galera-2" Dec 06 03:49:51 crc kubenswrapper[4980]: I1206 03:49:51.881544 4980 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/openstack-galera-0"] Dec 06 03:49:51 crc kubenswrapper[4980]: I1206 03:49:51.889464 4980 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/openstack-galera-1"] Dec 06 03:49:51 crc kubenswrapper[4980]: I1206 03:49:51.891808 4980 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/openstack-galera-2"] Dec 06 03:49:51 crc kubenswrapper[4980]: I1206 03:49:51.943925 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/7267968c-6dd4-44a8-ad87-ab5d0663d88a-config-data-generated\") pod \"openstack-galera-0\" (UID: \"7267968c-6dd4-44a8-ad87-ab5d0663d88a\") " pod="glance-kuttl-tests/openstack-galera-0" Dec 06 03:49:51 crc kubenswrapper[4980]: I1206 03:49:51.944009 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"openstack-galera-0\" (UID: \"7267968c-6dd4-44a8-ad87-ab5d0663d88a\") " pod="glance-kuttl-tests/openstack-galera-0" Dec 06 03:49:51 crc kubenswrapper[4980]: I1206 03:49:51.944191 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jbc2q\" (UniqueName: \"kubernetes.io/projected/7267968c-6dd4-44a8-ad87-ab5d0663d88a-kube-api-access-jbc2q\") pod \"openstack-galera-0\" (UID: \"7267968c-6dd4-44a8-ad87-ab5d0663d88a\") " pod="glance-kuttl-tests/openstack-galera-0" Dec 06 03:49:51 crc kubenswrapper[4980]: I1206 03:49:51.944214 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7267968c-6dd4-44a8-ad87-ab5d0663d88a-operator-scripts\") pod \"openstack-galera-0\" (UID: \"7267968c-6dd4-44a8-ad87-ab5d0663d88a\") " pod="glance-kuttl-tests/openstack-galera-0" Dec 06 03:49:51 crc kubenswrapper[4980]: I1206 03:49:51.944255 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/7267968c-6dd4-44a8-ad87-ab5d0663d88a-kolla-config\") pod \"openstack-galera-0\" (UID: \"7267968c-6dd4-44a8-ad87-ab5d0663d88a\") " pod="glance-kuttl-tests/openstack-galera-0" Dec 06 03:49:51 crc kubenswrapper[4980]: I1206 03:49:51.944282 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/7267968c-6dd4-44a8-ad87-ab5d0663d88a-config-data-default\") pod \"openstack-galera-0\" (UID: \"7267968c-6dd4-44a8-ad87-ab5d0663d88a\") " pod="glance-kuttl-tests/openstack-galera-0" Dec 06 03:49:51 crc kubenswrapper[4980]: I1206 03:49:51.944320 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/7267968c-6dd4-44a8-ad87-ab5d0663d88a-secrets\") pod \"openstack-galera-0\" (UID: \"7267968c-6dd4-44a8-ad87-ab5d0663d88a\") " pod="glance-kuttl-tests/openstack-galera-0" Dec 06 03:49:52 crc kubenswrapper[4980]: I1206 03:49:52.045205 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/c53b1fad-da29-4e9c-970c-9f4d84693905-kolla-config\") pod \"openstack-galera-2\" (UID: \"c53b1fad-da29-4e9c-970c-9f4d84693905\") " pod="glance-kuttl-tests/openstack-galera-2" Dec 06 03:49:52 crc kubenswrapper[4980]: I1206 03:49:52.045252 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/7267968c-6dd4-44a8-ad87-ab5d0663d88a-secrets\") pod \"openstack-galera-0\" (UID: \"7267968c-6dd4-44a8-ad87-ab5d0663d88a\") " pod="glance-kuttl-tests/openstack-galera-0" Dec 06 03:49:52 crc kubenswrapper[4980]: I1206 03:49:52.045273 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/c53b1fad-da29-4e9c-970c-9f4d84693905-secrets\") pod \"openstack-galera-2\" (UID: \"c53b1fad-da29-4e9c-970c-9f4d84693905\") " pod="glance-kuttl-tests/openstack-galera-2" Dec 06 03:49:52 crc kubenswrapper[4980]: I1206 03:49:52.045291 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/1f395fef-da33-429d-b7e0-61573b301d8d-secrets\") pod \"openstack-galera-1\" (UID: \"1f395fef-da33-429d-b7e0-61573b301d8d\") " pod="glance-kuttl-tests/openstack-galera-1" Dec 06 03:49:52 crc kubenswrapper[4980]: I1206 03:49:52.045344 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"openstack-galera-2\" (UID: \"c53b1fad-da29-4e9c-970c-9f4d84693905\") " pod="glance-kuttl-tests/openstack-galera-2" Dec 06 03:49:52 crc kubenswrapper[4980]: I1206 03:49:52.045401 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/c53b1fad-da29-4e9c-970c-9f4d84693905-config-data-generated\") pod \"openstack-galera-2\" (UID: \"c53b1fad-da29-4e9c-970c-9f4d84693905\") " pod="glance-kuttl-tests/openstack-galera-2" Dec 06 03:49:52 crc kubenswrapper[4980]: I1206 03:49:52.045430 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1f395fef-da33-429d-b7e0-61573b301d8d-operator-scripts\") pod \"openstack-galera-1\" (UID: \"1f395fef-da33-429d-b7e0-61573b301d8d\") " pod="glance-kuttl-tests/openstack-galera-1" Dec 06 03:49:52 crc kubenswrapper[4980]: I1206 03:49:52.045479 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/7267968c-6dd4-44a8-ad87-ab5d0663d88a-config-data-generated\") pod \"openstack-galera-0\" (UID: \"7267968c-6dd4-44a8-ad87-ab5d0663d88a\") " pod="glance-kuttl-tests/openstack-galera-0" Dec 06 03:49:52 crc kubenswrapper[4980]: I1206 03:49:52.045496 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vr5df\" (UniqueName: \"kubernetes.io/projected/c53b1fad-da29-4e9c-970c-9f4d84693905-kube-api-access-vr5df\") pod \"openstack-galera-2\" (UID: \"c53b1fad-da29-4e9c-970c-9f4d84693905\") " pod="glance-kuttl-tests/openstack-galera-2" Dec 06 03:49:52 crc kubenswrapper[4980]: I1206 03:49:52.045828 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/7267968c-6dd4-44a8-ad87-ab5d0663d88a-config-data-generated\") pod \"openstack-galera-0\" (UID: \"7267968c-6dd4-44a8-ad87-ab5d0663d88a\") " pod="glance-kuttl-tests/openstack-galera-0" Dec 06 03:49:52 crc kubenswrapper[4980]: I1206 03:49:52.045883 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6sgmz\" (UniqueName: \"kubernetes.io/projected/1f395fef-da33-429d-b7e0-61573b301d8d-kube-api-access-6sgmz\") pod \"openstack-galera-1\" (UID: \"1f395fef-da33-429d-b7e0-61573b301d8d\") " pod="glance-kuttl-tests/openstack-galera-1" Dec 06 03:49:52 crc kubenswrapper[4980]: I1206 03:49:52.045913 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"openstack-galera-0\" (UID: \"7267968c-6dd4-44a8-ad87-ab5d0663d88a\") " pod="glance-kuttl-tests/openstack-galera-0" Dec 06 03:49:52 crc kubenswrapper[4980]: I1206 03:49:52.045963 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jbc2q\" (UniqueName: \"kubernetes.io/projected/7267968c-6dd4-44a8-ad87-ab5d0663d88a-kube-api-access-jbc2q\") pod \"openstack-galera-0\" (UID: \"7267968c-6dd4-44a8-ad87-ab5d0663d88a\") " pod="glance-kuttl-tests/openstack-galera-0" Dec 06 03:49:52 crc kubenswrapper[4980]: I1206 03:49:52.045980 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7267968c-6dd4-44a8-ad87-ab5d0663d88a-operator-scripts\") pod \"openstack-galera-0\" (UID: \"7267968c-6dd4-44a8-ad87-ab5d0663d88a\") " pod="glance-kuttl-tests/openstack-galera-0" Dec 06 03:49:52 crc kubenswrapper[4980]: I1206 03:49:52.046646 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/1f395fef-da33-429d-b7e0-61573b301d8d-config-data-generated\") pod \"openstack-galera-1\" (UID: \"1f395fef-da33-429d-b7e0-61573b301d8d\") " pod="glance-kuttl-tests/openstack-galera-1" Dec 06 03:49:52 crc kubenswrapper[4980]: I1206 03:49:52.046698 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c53b1fad-da29-4e9c-970c-9f4d84693905-operator-scripts\") pod \"openstack-galera-2\" (UID: \"c53b1fad-da29-4e9c-970c-9f4d84693905\") " pod="glance-kuttl-tests/openstack-galera-2" Dec 06 03:49:52 crc kubenswrapper[4980]: I1206 03:49:52.046731 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"openstack-galera-1\" (UID: \"1f395fef-da33-429d-b7e0-61573b301d8d\") " pod="glance-kuttl-tests/openstack-galera-1" Dec 06 03:49:52 crc kubenswrapper[4980]: I1206 03:49:52.046759 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/1f395fef-da33-429d-b7e0-61573b301d8d-kolla-config\") pod \"openstack-galera-1\" (UID: \"1f395fef-da33-429d-b7e0-61573b301d8d\") " pod="glance-kuttl-tests/openstack-galera-1" Dec 06 03:49:52 crc kubenswrapper[4980]: I1206 03:49:52.046787 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/7267968c-6dd4-44a8-ad87-ab5d0663d88a-kolla-config\") pod \"openstack-galera-0\" (UID: \"7267968c-6dd4-44a8-ad87-ab5d0663d88a\") " pod="glance-kuttl-tests/openstack-galera-0" Dec 06 03:49:52 crc kubenswrapper[4980]: I1206 03:49:52.046814 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/7267968c-6dd4-44a8-ad87-ab5d0663d88a-config-data-default\") pod \"openstack-galera-0\" (UID: \"7267968c-6dd4-44a8-ad87-ab5d0663d88a\") " pod="glance-kuttl-tests/openstack-galera-0" Dec 06 03:49:52 crc kubenswrapper[4980]: I1206 03:49:52.046869 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/c53b1fad-da29-4e9c-970c-9f4d84693905-config-data-default\") pod \"openstack-galera-2\" (UID: \"c53b1fad-da29-4e9c-970c-9f4d84693905\") " pod="glance-kuttl-tests/openstack-galera-2" Dec 06 03:49:52 crc kubenswrapper[4980]: I1206 03:49:52.046894 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/1f395fef-da33-429d-b7e0-61573b301d8d-config-data-default\") pod \"openstack-galera-1\" (UID: \"1f395fef-da33-429d-b7e0-61573b301d8d\") " pod="glance-kuttl-tests/openstack-galera-1" Dec 06 03:49:52 crc kubenswrapper[4980]: I1206 03:49:52.046174 4980 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"openstack-galera-0\" (UID: \"7267968c-6dd4-44a8-ad87-ab5d0663d88a\") device mount path \"/mnt/openstack/pv02\"" pod="glance-kuttl-tests/openstack-galera-0" Dec 06 03:49:52 crc kubenswrapper[4980]: I1206 03:49:52.047281 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7267968c-6dd4-44a8-ad87-ab5d0663d88a-operator-scripts\") pod \"openstack-galera-0\" (UID: \"7267968c-6dd4-44a8-ad87-ab5d0663d88a\") " pod="glance-kuttl-tests/openstack-galera-0" Dec 06 03:49:52 crc kubenswrapper[4980]: I1206 03:49:52.047987 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/7267968c-6dd4-44a8-ad87-ab5d0663d88a-config-data-default\") pod \"openstack-galera-0\" (UID: \"7267968c-6dd4-44a8-ad87-ab5d0663d88a\") " pod="glance-kuttl-tests/openstack-galera-0" Dec 06 03:49:52 crc kubenswrapper[4980]: I1206 03:49:52.048379 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/7267968c-6dd4-44a8-ad87-ab5d0663d88a-kolla-config\") pod \"openstack-galera-0\" (UID: \"7267968c-6dd4-44a8-ad87-ab5d0663d88a\") " pod="glance-kuttl-tests/openstack-galera-0" Dec 06 03:49:52 crc kubenswrapper[4980]: I1206 03:49:52.058354 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/7267968c-6dd4-44a8-ad87-ab5d0663d88a-secrets\") pod \"openstack-galera-0\" (UID: \"7267968c-6dd4-44a8-ad87-ab5d0663d88a\") " pod="glance-kuttl-tests/openstack-galera-0" Dec 06 03:49:52 crc kubenswrapper[4980]: I1206 03:49:52.063292 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jbc2q\" (UniqueName: \"kubernetes.io/projected/7267968c-6dd4-44a8-ad87-ab5d0663d88a-kube-api-access-jbc2q\") pod \"openstack-galera-0\" (UID: \"7267968c-6dd4-44a8-ad87-ab5d0663d88a\") " pod="glance-kuttl-tests/openstack-galera-0" Dec 06 03:49:52 crc kubenswrapper[4980]: I1206 03:49:52.068058 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"openstack-galera-0\" (UID: \"7267968c-6dd4-44a8-ad87-ab5d0663d88a\") " pod="glance-kuttl-tests/openstack-galera-0" Dec 06 03:49:52 crc kubenswrapper[4980]: I1206 03:49:52.147824 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/c53b1fad-da29-4e9c-970c-9f4d84693905-config-data-default\") pod \"openstack-galera-2\" (UID: \"c53b1fad-da29-4e9c-970c-9f4d84693905\") " pod="glance-kuttl-tests/openstack-galera-2" Dec 06 03:49:52 crc kubenswrapper[4980]: I1206 03:49:52.147873 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/1f395fef-da33-429d-b7e0-61573b301d8d-config-data-default\") pod \"openstack-galera-1\" (UID: \"1f395fef-da33-429d-b7e0-61573b301d8d\") " pod="glance-kuttl-tests/openstack-galera-1" Dec 06 03:49:52 crc kubenswrapper[4980]: I1206 03:49:52.147899 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/c53b1fad-da29-4e9c-970c-9f4d84693905-kolla-config\") pod \"openstack-galera-2\" (UID: \"c53b1fad-da29-4e9c-970c-9f4d84693905\") " pod="glance-kuttl-tests/openstack-galera-2" Dec 06 03:49:52 crc kubenswrapper[4980]: I1206 03:49:52.147931 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/c53b1fad-da29-4e9c-970c-9f4d84693905-secrets\") pod \"openstack-galera-2\" (UID: \"c53b1fad-da29-4e9c-970c-9f4d84693905\") " pod="glance-kuttl-tests/openstack-galera-2" Dec 06 03:49:52 crc kubenswrapper[4980]: I1206 03:49:52.147954 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/1f395fef-da33-429d-b7e0-61573b301d8d-secrets\") pod \"openstack-galera-1\" (UID: \"1f395fef-da33-429d-b7e0-61573b301d8d\") " pod="glance-kuttl-tests/openstack-galera-1" Dec 06 03:49:52 crc kubenswrapper[4980]: I1206 03:49:52.147974 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"openstack-galera-2\" (UID: \"c53b1fad-da29-4e9c-970c-9f4d84693905\") " pod="glance-kuttl-tests/openstack-galera-2" Dec 06 03:49:52 crc kubenswrapper[4980]: I1206 03:49:52.147997 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/c53b1fad-da29-4e9c-970c-9f4d84693905-config-data-generated\") pod \"openstack-galera-2\" (UID: \"c53b1fad-da29-4e9c-970c-9f4d84693905\") " pod="glance-kuttl-tests/openstack-galera-2" Dec 06 03:49:52 crc kubenswrapper[4980]: I1206 03:49:52.148026 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1f395fef-da33-429d-b7e0-61573b301d8d-operator-scripts\") pod \"openstack-galera-1\" (UID: \"1f395fef-da33-429d-b7e0-61573b301d8d\") " pod="glance-kuttl-tests/openstack-galera-1" Dec 06 03:49:52 crc kubenswrapper[4980]: I1206 03:49:52.148050 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vr5df\" (UniqueName: \"kubernetes.io/projected/c53b1fad-da29-4e9c-970c-9f4d84693905-kube-api-access-vr5df\") pod \"openstack-galera-2\" (UID: \"c53b1fad-da29-4e9c-970c-9f4d84693905\") " pod="glance-kuttl-tests/openstack-galera-2" Dec 06 03:49:52 crc kubenswrapper[4980]: I1206 03:49:52.148072 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6sgmz\" (UniqueName: \"kubernetes.io/projected/1f395fef-da33-429d-b7e0-61573b301d8d-kube-api-access-6sgmz\") pod \"openstack-galera-1\" (UID: \"1f395fef-da33-429d-b7e0-61573b301d8d\") " pod="glance-kuttl-tests/openstack-galera-1" Dec 06 03:49:52 crc kubenswrapper[4980]: I1206 03:49:52.148116 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/1f395fef-da33-429d-b7e0-61573b301d8d-config-data-generated\") pod \"openstack-galera-1\" (UID: \"1f395fef-da33-429d-b7e0-61573b301d8d\") " pod="glance-kuttl-tests/openstack-galera-1" Dec 06 03:49:52 crc kubenswrapper[4980]: I1206 03:49:52.148149 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c53b1fad-da29-4e9c-970c-9f4d84693905-operator-scripts\") pod \"openstack-galera-2\" (UID: \"c53b1fad-da29-4e9c-970c-9f4d84693905\") " pod="glance-kuttl-tests/openstack-galera-2" Dec 06 03:49:52 crc kubenswrapper[4980]: I1206 03:49:52.148194 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"openstack-galera-1\" (UID: \"1f395fef-da33-429d-b7e0-61573b301d8d\") " pod="glance-kuttl-tests/openstack-galera-1" Dec 06 03:49:52 crc kubenswrapper[4980]: I1206 03:49:52.148222 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/1f395fef-da33-429d-b7e0-61573b301d8d-kolla-config\") pod \"openstack-galera-1\" (UID: \"1f395fef-da33-429d-b7e0-61573b301d8d\") " pod="glance-kuttl-tests/openstack-galera-1" Dec 06 03:49:52 crc kubenswrapper[4980]: I1206 03:49:52.148851 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/c53b1fad-da29-4e9c-970c-9f4d84693905-kolla-config\") pod \"openstack-galera-2\" (UID: \"c53b1fad-da29-4e9c-970c-9f4d84693905\") " pod="glance-kuttl-tests/openstack-galera-2" Dec 06 03:49:52 crc kubenswrapper[4980]: I1206 03:49:52.149200 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/1f395fef-da33-429d-b7e0-61573b301d8d-kolla-config\") pod \"openstack-galera-1\" (UID: \"1f395fef-da33-429d-b7e0-61573b301d8d\") " pod="glance-kuttl-tests/openstack-galera-1" Dec 06 03:49:52 crc kubenswrapper[4980]: I1206 03:49:52.149563 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/1f395fef-da33-429d-b7e0-61573b301d8d-config-data-default\") pod \"openstack-galera-1\" (UID: \"1f395fef-da33-429d-b7e0-61573b301d8d\") " pod="glance-kuttl-tests/openstack-galera-1" Dec 06 03:49:52 crc kubenswrapper[4980]: I1206 03:49:52.149714 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/c53b1fad-da29-4e9c-970c-9f4d84693905-config-data-generated\") pod \"openstack-galera-2\" (UID: \"c53b1fad-da29-4e9c-970c-9f4d84693905\") " pod="glance-kuttl-tests/openstack-galera-2" Dec 06 03:49:52 crc kubenswrapper[4980]: I1206 03:49:52.149905 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/1f395fef-da33-429d-b7e0-61573b301d8d-config-data-generated\") pod \"openstack-galera-1\" (UID: \"1f395fef-da33-429d-b7e0-61573b301d8d\") " pod="glance-kuttl-tests/openstack-galera-1" Dec 06 03:49:52 crc kubenswrapper[4980]: I1206 03:49:52.156327 4980 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"openstack-galera-1\" (UID: \"1f395fef-da33-429d-b7e0-61573b301d8d\") device mount path \"/mnt/openstack/pv03\"" pod="glance-kuttl-tests/openstack-galera-1" Dec 06 03:49:52 crc kubenswrapper[4980]: I1206 03:49:52.156413 4980 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"openstack-galera-2\" (UID: \"c53b1fad-da29-4e9c-970c-9f4d84693905\") device mount path \"/mnt/openstack/pv04\"" pod="glance-kuttl-tests/openstack-galera-2" Dec 06 03:49:52 crc kubenswrapper[4980]: I1206 03:49:52.156543 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/1f395fef-da33-429d-b7e0-61573b301d8d-secrets\") pod \"openstack-galera-1\" (UID: \"1f395fef-da33-429d-b7e0-61573b301d8d\") " pod="glance-kuttl-tests/openstack-galera-1" Dec 06 03:49:52 crc kubenswrapper[4980]: I1206 03:49:52.157029 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/c53b1fad-da29-4e9c-970c-9f4d84693905-config-data-default\") pod \"openstack-galera-2\" (UID: \"c53b1fad-da29-4e9c-970c-9f4d84693905\") " pod="glance-kuttl-tests/openstack-galera-2" Dec 06 03:49:52 crc kubenswrapper[4980]: I1206 03:49:52.157433 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1f395fef-da33-429d-b7e0-61573b301d8d-operator-scripts\") pod \"openstack-galera-1\" (UID: \"1f395fef-da33-429d-b7e0-61573b301d8d\") " pod="glance-kuttl-tests/openstack-galera-1" Dec 06 03:49:52 crc kubenswrapper[4980]: I1206 03:49:52.172549 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/openstack-galera-0" Dec 06 03:49:52 crc kubenswrapper[4980]: I1206 03:49:52.173601 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c53b1fad-da29-4e9c-970c-9f4d84693905-operator-scripts\") pod \"openstack-galera-2\" (UID: \"c53b1fad-da29-4e9c-970c-9f4d84693905\") " pod="glance-kuttl-tests/openstack-galera-2" Dec 06 03:49:52 crc kubenswrapper[4980]: I1206 03:49:52.174525 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/c53b1fad-da29-4e9c-970c-9f4d84693905-secrets\") pod \"openstack-galera-2\" (UID: \"c53b1fad-da29-4e9c-970c-9f4d84693905\") " pod="glance-kuttl-tests/openstack-galera-2" Dec 06 03:49:52 crc kubenswrapper[4980]: I1206 03:49:52.174781 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6sgmz\" (UniqueName: \"kubernetes.io/projected/1f395fef-da33-429d-b7e0-61573b301d8d-kube-api-access-6sgmz\") pod \"openstack-galera-1\" (UID: \"1f395fef-da33-429d-b7e0-61573b301d8d\") " pod="glance-kuttl-tests/openstack-galera-1" Dec 06 03:49:52 crc kubenswrapper[4980]: I1206 03:49:52.178118 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vr5df\" (UniqueName: \"kubernetes.io/projected/c53b1fad-da29-4e9c-970c-9f4d84693905-kube-api-access-vr5df\") pod \"openstack-galera-2\" (UID: \"c53b1fad-da29-4e9c-970c-9f4d84693905\") " pod="glance-kuttl-tests/openstack-galera-2" Dec 06 03:49:52 crc kubenswrapper[4980]: I1206 03:49:52.190696 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"openstack-galera-2\" (UID: \"c53b1fad-da29-4e9c-970c-9f4d84693905\") " pod="glance-kuttl-tests/openstack-galera-2" Dec 06 03:49:52 crc kubenswrapper[4980]: I1206 03:49:52.196714 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"openstack-galera-1\" (UID: \"1f395fef-da33-429d-b7e0-61573b301d8d\") " pod="glance-kuttl-tests/openstack-galera-1" Dec 06 03:49:52 crc kubenswrapper[4980]: I1206 03:49:52.204227 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/openstack-galera-1" Dec 06 03:49:52 crc kubenswrapper[4980]: I1206 03:49:52.210374 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/openstack-galera-2" Dec 06 03:49:52 crc kubenswrapper[4980]: I1206 03:49:52.690859 4980 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/openstack-galera-1"] Dec 06 03:49:52 crc kubenswrapper[4980]: W1206 03:49:52.694636 4980 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1f395fef_da33_429d_b7e0_61573b301d8d.slice/crio-3fd2af2e4daff1c483abdf0474ec080df95194afcff5a053b3cf9179b98c19f8 WatchSource:0}: Error finding container 3fd2af2e4daff1c483abdf0474ec080df95194afcff5a053b3cf9179b98c19f8: Status 404 returned error can't find the container with id 3fd2af2e4daff1c483abdf0474ec080df95194afcff5a053b3cf9179b98c19f8 Dec 06 03:49:52 crc kubenswrapper[4980]: I1206 03:49:52.746436 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/openstack-galera-1" event={"ID":"1f395fef-da33-429d-b7e0-61573b301d8d","Type":"ContainerStarted","Data":"3fd2af2e4daff1c483abdf0474ec080df95194afcff5a053b3cf9179b98c19f8"} Dec 06 03:49:52 crc kubenswrapper[4980]: I1206 03:49:52.747280 4980 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/openstack-galera-0"] Dec 06 03:49:53 crc kubenswrapper[4980]: I1206 03:49:53.013792 4980 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/openstack-galera-2"] Dec 06 03:49:53 crc kubenswrapper[4980]: W1206 03:49:53.017577 4980 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc53b1fad_da29_4e9c_970c_9f4d84693905.slice/crio-7befaa8068352e7550da1bd63bc17cadffc803c0463080341725f67e0e6d9336 WatchSource:0}: Error finding container 7befaa8068352e7550da1bd63bc17cadffc803c0463080341725f67e0e6d9336: Status 404 returned error can't find the container with id 7befaa8068352e7550da1bd63bc17cadffc803c0463080341725f67e0e6d9336 Dec 06 03:49:53 crc kubenswrapper[4980]: I1206 03:49:53.753288 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-99587f5d7-rcrhn" event={"ID":"0cc75ddc-77ef-4c34-b7d1-c5e98f7666dc","Type":"ContainerStarted","Data":"7b62393371ddebc8709da8fc3544208451022933ae5f4eca89d4562ea7efcc70"} Dec 06 03:49:53 crc kubenswrapper[4980]: I1206 03:49:53.753356 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-99587f5d7-rcrhn" event={"ID":"0cc75ddc-77ef-4c34-b7d1-c5e98f7666dc","Type":"ContainerStarted","Data":"f8fdba35fb6deedf9e3600be7ba0f0499b8c223c3f6050217b524093e6a884ac"} Dec 06 03:49:53 crc kubenswrapper[4980]: I1206 03:49:53.753397 4980 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/infra-operator-controller-manager-99587f5d7-rcrhn" Dec 06 03:49:53 crc kubenswrapper[4980]: I1206 03:49:53.754706 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/openstack-galera-2" event={"ID":"c53b1fad-da29-4e9c-970c-9f4d84693905","Type":"ContainerStarted","Data":"7befaa8068352e7550da1bd63bc17cadffc803c0463080341725f67e0e6d9336"} Dec 06 03:49:53 crc kubenswrapper[4980]: I1206 03:49:53.756032 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/openstack-galera-0" event={"ID":"7267968c-6dd4-44a8-ad87-ab5d0663d88a","Type":"ContainerStarted","Data":"f36fcc05f71495d110a756f21ca79e4cbc5c7bc2c8148d9798aee5eaf7d1550b"} Dec 06 03:49:53 crc kubenswrapper[4980]: I1206 03:49:53.774918 4980 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/infra-operator-controller-manager-99587f5d7-rcrhn" podStartSLOduration=1.879352023 podStartE2EDuration="3.774851794s" podCreationTimestamp="2025-12-06 03:49:50 +0000 UTC" firstStartedPulling="2025-12-06 03:49:50.599843541 +0000 UTC m=+949.844594812" lastFinishedPulling="2025-12-06 03:49:52.495343312 +0000 UTC m=+951.740094583" observedRunningTime="2025-12-06 03:49:53.770253381 +0000 UTC m=+953.015004672" watchObservedRunningTime="2025-12-06 03:49:53.774851794 +0000 UTC m=+953.019603065" Dec 06 03:49:55 crc kubenswrapper[4980]: I1206 03:49:55.137468 4980 patch_prober.go:28] interesting pod/machine-config-daemon-r5zfb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 03:49:55 crc kubenswrapper[4980]: I1206 03:49:55.137799 4980 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-r5zfb" podUID="320f44d1-a671-4a91-b328-a8b0fdd8f23a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 03:50:00 crc kubenswrapper[4980]: I1206 03:50:00.396292 4980 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/infra-operator-controller-manager-99587f5d7-rcrhn" Dec 06 03:50:06 crc kubenswrapper[4980]: I1206 03:50:06.858888 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/openstack-galera-1" event={"ID":"1f395fef-da33-429d-b7e0-61573b301d8d","Type":"ContainerStarted","Data":"f158e5bfc81aaf82fb16897f19b40d4701f26b887726439164ad87e134f4efce"} Dec 06 03:50:06 crc kubenswrapper[4980]: I1206 03:50:06.860484 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/openstack-galera-2" event={"ID":"c53b1fad-da29-4e9c-970c-9f4d84693905","Type":"ContainerStarted","Data":"2c7247d3e3e5f9be36e1b2393fa78b87e248e8f6fde1e6bc618d86f5de98c1a2"} Dec 06 03:50:06 crc kubenswrapper[4980]: I1206 03:50:06.861394 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/openstack-galera-0" event={"ID":"7267968c-6dd4-44a8-ad87-ab5d0663d88a","Type":"ContainerStarted","Data":"72ae2effc98f017bcd92aadbd65b11aa3bba6562e150bde4c5a8d8d7bb2cb319"} Dec 06 03:50:09 crc kubenswrapper[4980]: I1206 03:50:09.037322 4980 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-index-fw7cs"] Dec 06 03:50:09 crc kubenswrapper[4980]: I1206 03:50:09.038129 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-index-fw7cs" Dec 06 03:50:09 crc kubenswrapper[4980]: I1206 03:50:09.040613 4980 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"rabbitmq-cluster-operator-index-dockercfg-nfmlx" Dec 06 03:50:09 crc kubenswrapper[4980]: I1206 03:50:09.046142 4980 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-index-fw7cs"] Dec 06 03:50:09 crc kubenswrapper[4980]: I1206 03:50:09.110455 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jh5ww\" (UniqueName: \"kubernetes.io/projected/71de7a2d-0537-494a-be72-446c4be812e5-kube-api-access-jh5ww\") pod \"rabbitmq-cluster-operator-index-fw7cs\" (UID: \"71de7a2d-0537-494a-be72-446c4be812e5\") " pod="openstack-operators/rabbitmq-cluster-operator-index-fw7cs" Dec 06 03:50:09 crc kubenswrapper[4980]: I1206 03:50:09.211623 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jh5ww\" (UniqueName: \"kubernetes.io/projected/71de7a2d-0537-494a-be72-446c4be812e5-kube-api-access-jh5ww\") pod \"rabbitmq-cluster-operator-index-fw7cs\" (UID: \"71de7a2d-0537-494a-be72-446c4be812e5\") " pod="openstack-operators/rabbitmq-cluster-operator-index-fw7cs" Dec 06 03:50:09 crc kubenswrapper[4980]: I1206 03:50:09.230962 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jh5ww\" (UniqueName: \"kubernetes.io/projected/71de7a2d-0537-494a-be72-446c4be812e5-kube-api-access-jh5ww\") pod \"rabbitmq-cluster-operator-index-fw7cs\" (UID: \"71de7a2d-0537-494a-be72-446c4be812e5\") " pod="openstack-operators/rabbitmq-cluster-operator-index-fw7cs" Dec 06 03:50:09 crc kubenswrapper[4980]: I1206 03:50:09.355394 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-index-fw7cs" Dec 06 03:50:09 crc kubenswrapper[4980]: I1206 03:50:09.791928 4980 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-index-fw7cs"] Dec 06 03:50:09 crc kubenswrapper[4980]: I1206 03:50:09.877725 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-index-fw7cs" event={"ID":"71de7a2d-0537-494a-be72-446c4be812e5","Type":"ContainerStarted","Data":"2fe15c36b762f5f696746bb7656027d378497ddbf97fd211f3f87271db8f7a6d"} Dec 06 03:50:11 crc kubenswrapper[4980]: I1206 03:50:11.892111 4980 generic.go:334] "Generic (PLEG): container finished" podID="7267968c-6dd4-44a8-ad87-ab5d0663d88a" containerID="72ae2effc98f017bcd92aadbd65b11aa3bba6562e150bde4c5a8d8d7bb2cb319" exitCode=0 Dec 06 03:50:11 crc kubenswrapper[4980]: I1206 03:50:11.892195 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/openstack-galera-0" event={"ID":"7267968c-6dd4-44a8-ad87-ab5d0663d88a","Type":"ContainerDied","Data":"72ae2effc98f017bcd92aadbd65b11aa3bba6562e150bde4c5a8d8d7bb2cb319"} Dec 06 03:50:11 crc kubenswrapper[4980]: I1206 03:50:11.894802 4980 generic.go:334] "Generic (PLEG): container finished" podID="1f395fef-da33-429d-b7e0-61573b301d8d" containerID="f158e5bfc81aaf82fb16897f19b40d4701f26b887726439164ad87e134f4efce" exitCode=0 Dec 06 03:50:11 crc kubenswrapper[4980]: I1206 03:50:11.894843 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/openstack-galera-1" event={"ID":"1f395fef-da33-429d-b7e0-61573b301d8d","Type":"ContainerDied","Data":"f158e5bfc81aaf82fb16897f19b40d4701f26b887726439164ad87e134f4efce"} Dec 06 03:50:11 crc kubenswrapper[4980]: I1206 03:50:11.897234 4980 generic.go:334] "Generic (PLEG): container finished" podID="c53b1fad-da29-4e9c-970c-9f4d84693905" containerID="2c7247d3e3e5f9be36e1b2393fa78b87e248e8f6fde1e6bc618d86f5de98c1a2" exitCode=0 Dec 06 03:50:11 crc kubenswrapper[4980]: I1206 03:50:11.897265 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/openstack-galera-2" event={"ID":"c53b1fad-da29-4e9c-970c-9f4d84693905","Type":"ContainerDied","Data":"2c7247d3e3e5f9be36e1b2393fa78b87e248e8f6fde1e6bc618d86f5de98c1a2"} Dec 06 03:50:13 crc kubenswrapper[4980]: I1206 03:50:13.421777 4980 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-index-fw7cs"] Dec 06 03:50:13 crc kubenswrapper[4980]: I1206 03:50:13.919358 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-index-fw7cs" event={"ID":"71de7a2d-0537-494a-be72-446c4be812e5","Type":"ContainerStarted","Data":"c0b28253664881c9a29de28465fe26c5e50947318e47c718e96c903532027c93"} Dec 06 03:50:13 crc kubenswrapper[4980]: I1206 03:50:13.919551 4980 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/rabbitmq-cluster-operator-index-fw7cs" podUID="71de7a2d-0537-494a-be72-446c4be812e5" containerName="registry-server" containerID="cri-o://c0b28253664881c9a29de28465fe26c5e50947318e47c718e96c903532027c93" gracePeriod=2 Dec 06 03:50:13 crc kubenswrapper[4980]: I1206 03:50:13.921981 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/openstack-galera-2" event={"ID":"c53b1fad-da29-4e9c-970c-9f4d84693905","Type":"ContainerStarted","Data":"ff80dcb9df94d71d9b7bbe3d1a4cfb752e5385542b0e6c2998b5e462a303a3dd"} Dec 06 03:50:13 crc kubenswrapper[4980]: I1206 03:50:13.924753 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/openstack-galera-0" event={"ID":"7267968c-6dd4-44a8-ad87-ab5d0663d88a","Type":"ContainerStarted","Data":"74a9442bb3e0e40fc5b7df44bb1913e318f991469071e3ec08cb0e5665e13f54"} Dec 06 03:50:13 crc kubenswrapper[4980]: I1206 03:50:13.928022 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/openstack-galera-1" event={"ID":"1f395fef-da33-429d-b7e0-61573b301d8d","Type":"ContainerStarted","Data":"938f31fa1632dbf3a95925e05e2932fee706e073253062b66ec5120058b8bda0"} Dec 06 03:50:13 crc kubenswrapper[4980]: I1206 03:50:13.941163 4980 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/rabbitmq-cluster-operator-index-fw7cs" podStartSLOduration=1.1396587 podStartE2EDuration="4.941136211s" podCreationTimestamp="2025-12-06 03:50:09 +0000 UTC" firstStartedPulling="2025-12-06 03:50:09.803366926 +0000 UTC m=+969.048118197" lastFinishedPulling="2025-12-06 03:50:13.604844437 +0000 UTC m=+972.849595708" observedRunningTime="2025-12-06 03:50:13.937187907 +0000 UTC m=+973.181939188" watchObservedRunningTime="2025-12-06 03:50:13.941136211 +0000 UTC m=+973.185887472" Dec 06 03:50:13 crc kubenswrapper[4980]: I1206 03:50:13.959313 4980 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/openstack-galera-1" podStartSLOduration=10.664772275 podStartE2EDuration="23.959288813s" podCreationTimestamp="2025-12-06 03:49:50 +0000 UTC" firstStartedPulling="2025-12-06 03:49:52.70573446 +0000 UTC m=+951.950485731" lastFinishedPulling="2025-12-06 03:50:06.000250988 +0000 UTC m=+965.245002269" observedRunningTime="2025-12-06 03:50:13.956556705 +0000 UTC m=+973.201307986" watchObservedRunningTime="2025-12-06 03:50:13.959288813 +0000 UTC m=+973.204040084" Dec 06 03:50:13 crc kubenswrapper[4980]: I1206 03:50:13.975591 4980 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/openstack-galera-0" podStartSLOduration=10.682632259 podStartE2EDuration="23.975563452s" podCreationTimestamp="2025-12-06 03:49:50 +0000 UTC" firstStartedPulling="2025-12-06 03:49:52.75611022 +0000 UTC m=+952.000861491" lastFinishedPulling="2025-12-06 03:50:06.049041413 +0000 UTC m=+965.293792684" observedRunningTime="2025-12-06 03:50:13.971611988 +0000 UTC m=+973.216363269" watchObservedRunningTime="2025-12-06 03:50:13.975563452 +0000 UTC m=+973.220314723" Dec 06 03:50:14 crc kubenswrapper[4980]: I1206 03:50:14.012359 4980 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/openstack-galera-2" podStartSLOduration=11.037030085 podStartE2EDuration="24.012342721s" podCreationTimestamp="2025-12-06 03:49:50 +0000 UTC" firstStartedPulling="2025-12-06 03:49:53.019163715 +0000 UTC m=+952.263914986" lastFinishedPulling="2025-12-06 03:50:05.994476341 +0000 UTC m=+965.239227622" observedRunningTime="2025-12-06 03:50:14.009684135 +0000 UTC m=+973.254435426" watchObservedRunningTime="2025-12-06 03:50:14.012342721 +0000 UTC m=+973.257093992" Dec 06 03:50:14 crc kubenswrapper[4980]: I1206 03:50:14.031905 4980 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-index-7m77w"] Dec 06 03:50:14 crc kubenswrapper[4980]: I1206 03:50:14.033607 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-index-7m77w" Dec 06 03:50:14 crc kubenswrapper[4980]: I1206 03:50:14.040268 4980 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-index-7m77w"] Dec 06 03:50:14 crc kubenswrapper[4980]: I1206 03:50:14.095634 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-98nh8\" (UniqueName: \"kubernetes.io/projected/b514140e-e078-40e1-b380-5c097450eeac-kube-api-access-98nh8\") pod \"rabbitmq-cluster-operator-index-7m77w\" (UID: \"b514140e-e078-40e1-b380-5c097450eeac\") " pod="openstack-operators/rabbitmq-cluster-operator-index-7m77w" Dec 06 03:50:14 crc kubenswrapper[4980]: I1206 03:50:14.197377 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-98nh8\" (UniqueName: \"kubernetes.io/projected/b514140e-e078-40e1-b380-5c097450eeac-kube-api-access-98nh8\") pod \"rabbitmq-cluster-operator-index-7m77w\" (UID: \"b514140e-e078-40e1-b380-5c097450eeac\") " pod="openstack-operators/rabbitmq-cluster-operator-index-7m77w" Dec 06 03:50:14 crc kubenswrapper[4980]: I1206 03:50:14.235203 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-98nh8\" (UniqueName: \"kubernetes.io/projected/b514140e-e078-40e1-b380-5c097450eeac-kube-api-access-98nh8\") pod \"rabbitmq-cluster-operator-index-7m77w\" (UID: \"b514140e-e078-40e1-b380-5c097450eeac\") " pod="openstack-operators/rabbitmq-cluster-operator-index-7m77w" Dec 06 03:50:14 crc kubenswrapper[4980]: I1206 03:50:14.286440 4980 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-index-fw7cs" Dec 06 03:50:14 crc kubenswrapper[4980]: I1206 03:50:14.387464 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-index-7m77w" Dec 06 03:50:14 crc kubenswrapper[4980]: I1206 03:50:14.400106 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jh5ww\" (UniqueName: \"kubernetes.io/projected/71de7a2d-0537-494a-be72-446c4be812e5-kube-api-access-jh5ww\") pod \"71de7a2d-0537-494a-be72-446c4be812e5\" (UID: \"71de7a2d-0537-494a-be72-446c4be812e5\") " Dec 06 03:50:14 crc kubenswrapper[4980]: I1206 03:50:14.403080 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/71de7a2d-0537-494a-be72-446c4be812e5-kube-api-access-jh5ww" (OuterVolumeSpecName: "kube-api-access-jh5ww") pod "71de7a2d-0537-494a-be72-446c4be812e5" (UID: "71de7a2d-0537-494a-be72-446c4be812e5"). InnerVolumeSpecName "kube-api-access-jh5ww". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 03:50:14 crc kubenswrapper[4980]: I1206 03:50:14.501394 4980 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jh5ww\" (UniqueName: \"kubernetes.io/projected/71de7a2d-0537-494a-be72-446c4be812e5-kube-api-access-jh5ww\") on node \"crc\" DevicePath \"\"" Dec 06 03:50:14 crc kubenswrapper[4980]: I1206 03:50:14.771782 4980 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-index-7m77w"] Dec 06 03:50:14 crc kubenswrapper[4980]: W1206 03:50:14.780631 4980 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb514140e_e078_40e1_b380_5c097450eeac.slice/crio-f42b1e24a9eba5a1a3af7bd012256245451ba29e1d9672dc9093b4916cccea73 WatchSource:0}: Error finding container f42b1e24a9eba5a1a3af7bd012256245451ba29e1d9672dc9093b4916cccea73: Status 404 returned error can't find the container with id f42b1e24a9eba5a1a3af7bd012256245451ba29e1d9672dc9093b4916cccea73 Dec 06 03:50:14 crc kubenswrapper[4980]: I1206 03:50:14.935029 4980 generic.go:334] "Generic (PLEG): container finished" podID="71de7a2d-0537-494a-be72-446c4be812e5" containerID="c0b28253664881c9a29de28465fe26c5e50947318e47c718e96c903532027c93" exitCode=0 Dec 06 03:50:14 crc kubenswrapper[4980]: I1206 03:50:14.935094 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-index-fw7cs" event={"ID":"71de7a2d-0537-494a-be72-446c4be812e5","Type":"ContainerDied","Data":"c0b28253664881c9a29de28465fe26c5e50947318e47c718e96c903532027c93"} Dec 06 03:50:14 crc kubenswrapper[4980]: I1206 03:50:14.935135 4980 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-index-fw7cs" Dec 06 03:50:14 crc kubenswrapper[4980]: I1206 03:50:14.935169 4980 scope.go:117] "RemoveContainer" containerID="c0b28253664881c9a29de28465fe26c5e50947318e47c718e96c903532027c93" Dec 06 03:50:14 crc kubenswrapper[4980]: I1206 03:50:14.935137 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-index-fw7cs" event={"ID":"71de7a2d-0537-494a-be72-446c4be812e5","Type":"ContainerDied","Data":"2fe15c36b762f5f696746bb7656027d378497ddbf97fd211f3f87271db8f7a6d"} Dec 06 03:50:14 crc kubenswrapper[4980]: I1206 03:50:14.936175 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-index-7m77w" event={"ID":"b514140e-e078-40e1-b380-5c097450eeac","Type":"ContainerStarted","Data":"f42b1e24a9eba5a1a3af7bd012256245451ba29e1d9672dc9093b4916cccea73"} Dec 06 03:50:14 crc kubenswrapper[4980]: I1206 03:50:14.952817 4980 scope.go:117] "RemoveContainer" containerID="c0b28253664881c9a29de28465fe26c5e50947318e47c718e96c903532027c93" Dec 06 03:50:14 crc kubenswrapper[4980]: E1206 03:50:14.953395 4980 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c0b28253664881c9a29de28465fe26c5e50947318e47c718e96c903532027c93\": container with ID starting with c0b28253664881c9a29de28465fe26c5e50947318e47c718e96c903532027c93 not found: ID does not exist" containerID="c0b28253664881c9a29de28465fe26c5e50947318e47c718e96c903532027c93" Dec 06 03:50:14 crc kubenswrapper[4980]: I1206 03:50:14.953431 4980 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c0b28253664881c9a29de28465fe26c5e50947318e47c718e96c903532027c93"} err="failed to get container status \"c0b28253664881c9a29de28465fe26c5e50947318e47c718e96c903532027c93\": rpc error: code = NotFound desc = could not find container \"c0b28253664881c9a29de28465fe26c5e50947318e47c718e96c903532027c93\": container with ID starting with c0b28253664881c9a29de28465fe26c5e50947318e47c718e96c903532027c93 not found: ID does not exist" Dec 06 03:50:14 crc kubenswrapper[4980]: I1206 03:50:14.959650 4980 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-index-fw7cs"] Dec 06 03:50:14 crc kubenswrapper[4980]: I1206 03:50:14.965796 4980 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-index-fw7cs"] Dec 06 03:50:15 crc kubenswrapper[4980]: I1206 03:50:15.196576 4980 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="71de7a2d-0537-494a-be72-446c4be812e5" path="/var/lib/kubelet/pods/71de7a2d-0537-494a-be72-446c4be812e5/volumes" Dec 06 03:50:15 crc kubenswrapper[4980]: I1206 03:50:15.946550 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-index-7m77w" event={"ID":"b514140e-e078-40e1-b380-5c097450eeac","Type":"ContainerStarted","Data":"378b56d81963654789e60c156b26acdbe238a082d2b5ad4534c144be0e8dfbe6"} Dec 06 03:50:15 crc kubenswrapper[4980]: I1206 03:50:15.977115 4980 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/rabbitmq-cluster-operator-index-7m77w" podStartSLOduration=1.564019811 podStartE2EDuration="1.977073084s" podCreationTimestamp="2025-12-06 03:50:14 +0000 UTC" firstStartedPulling="2025-12-06 03:50:14.787152711 +0000 UTC m=+974.031903982" lastFinishedPulling="2025-12-06 03:50:15.200205984 +0000 UTC m=+974.444957255" observedRunningTime="2025-12-06 03:50:15.966901551 +0000 UTC m=+975.211652862" watchObservedRunningTime="2025-12-06 03:50:15.977073084 +0000 UTC m=+975.221824445" Dec 06 03:50:16 crc kubenswrapper[4980]: I1206 03:50:16.450904 4980 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/memcached-0"] Dec 06 03:50:16 crc kubenswrapper[4980]: E1206 03:50:16.451227 4980 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="71de7a2d-0537-494a-be72-446c4be812e5" containerName="registry-server" Dec 06 03:50:16 crc kubenswrapper[4980]: I1206 03:50:16.451248 4980 state_mem.go:107] "Deleted CPUSet assignment" podUID="71de7a2d-0537-494a-be72-446c4be812e5" containerName="registry-server" Dec 06 03:50:16 crc kubenswrapper[4980]: I1206 03:50:16.451383 4980 memory_manager.go:354] "RemoveStaleState removing state" podUID="71de7a2d-0537-494a-be72-446c4be812e5" containerName="registry-server" Dec 06 03:50:16 crc kubenswrapper[4980]: I1206 03:50:16.451931 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/memcached-0" Dec 06 03:50:16 crc kubenswrapper[4980]: I1206 03:50:16.453791 4980 reflector.go:368] Caches populated for *v1.ConfigMap from object-"glance-kuttl-tests"/"memcached-config-data" Dec 06 03:50:16 crc kubenswrapper[4980]: I1206 03:50:16.454334 4980 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"memcached-memcached-dockercfg-4mg2j" Dec 06 03:50:16 crc kubenswrapper[4980]: I1206 03:50:16.463169 4980 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/memcached-0"] Dec 06 03:50:16 crc kubenswrapper[4980]: I1206 03:50:16.527637 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/cbbe1b5d-848b-4c0f-bd83-10b27906dbfc-kolla-config\") pod \"memcached-0\" (UID: \"cbbe1b5d-848b-4c0f-bd83-10b27906dbfc\") " pod="glance-kuttl-tests/memcached-0" Dec 06 03:50:16 crc kubenswrapper[4980]: I1206 03:50:16.527735 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/cbbe1b5d-848b-4c0f-bd83-10b27906dbfc-config-data\") pod \"memcached-0\" (UID: \"cbbe1b5d-848b-4c0f-bd83-10b27906dbfc\") " pod="glance-kuttl-tests/memcached-0" Dec 06 03:50:16 crc kubenswrapper[4980]: I1206 03:50:16.527765 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5mcdc\" (UniqueName: \"kubernetes.io/projected/cbbe1b5d-848b-4c0f-bd83-10b27906dbfc-kube-api-access-5mcdc\") pod \"memcached-0\" (UID: \"cbbe1b5d-848b-4c0f-bd83-10b27906dbfc\") " pod="glance-kuttl-tests/memcached-0" Dec 06 03:50:16 crc kubenswrapper[4980]: I1206 03:50:16.629404 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5mcdc\" (UniqueName: \"kubernetes.io/projected/cbbe1b5d-848b-4c0f-bd83-10b27906dbfc-kube-api-access-5mcdc\") pod \"memcached-0\" (UID: \"cbbe1b5d-848b-4c0f-bd83-10b27906dbfc\") " pod="glance-kuttl-tests/memcached-0" Dec 06 03:50:16 crc kubenswrapper[4980]: I1206 03:50:16.629529 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/cbbe1b5d-848b-4c0f-bd83-10b27906dbfc-kolla-config\") pod \"memcached-0\" (UID: \"cbbe1b5d-848b-4c0f-bd83-10b27906dbfc\") " pod="glance-kuttl-tests/memcached-0" Dec 06 03:50:16 crc kubenswrapper[4980]: I1206 03:50:16.629600 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/cbbe1b5d-848b-4c0f-bd83-10b27906dbfc-config-data\") pod \"memcached-0\" (UID: \"cbbe1b5d-848b-4c0f-bd83-10b27906dbfc\") " pod="glance-kuttl-tests/memcached-0" Dec 06 03:50:16 crc kubenswrapper[4980]: I1206 03:50:16.630467 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/cbbe1b5d-848b-4c0f-bd83-10b27906dbfc-kolla-config\") pod \"memcached-0\" (UID: \"cbbe1b5d-848b-4c0f-bd83-10b27906dbfc\") " pod="glance-kuttl-tests/memcached-0" Dec 06 03:50:16 crc kubenswrapper[4980]: I1206 03:50:16.631020 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/cbbe1b5d-848b-4c0f-bd83-10b27906dbfc-config-data\") pod \"memcached-0\" (UID: \"cbbe1b5d-848b-4c0f-bd83-10b27906dbfc\") " pod="glance-kuttl-tests/memcached-0" Dec 06 03:50:16 crc kubenswrapper[4980]: I1206 03:50:16.658585 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5mcdc\" (UniqueName: \"kubernetes.io/projected/cbbe1b5d-848b-4c0f-bd83-10b27906dbfc-kube-api-access-5mcdc\") pod \"memcached-0\" (UID: \"cbbe1b5d-848b-4c0f-bd83-10b27906dbfc\") " pod="glance-kuttl-tests/memcached-0" Dec 06 03:50:16 crc kubenswrapper[4980]: I1206 03:50:16.807822 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/memcached-0" Dec 06 03:50:17 crc kubenswrapper[4980]: I1206 03:50:17.224811 4980 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/memcached-0"] Dec 06 03:50:17 crc kubenswrapper[4980]: W1206 03:50:17.230748 4980 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podcbbe1b5d_848b_4c0f_bd83_10b27906dbfc.slice/crio-3bffb91d0bd950071f058ad2d50254e43e9cab2b2dc8f73ae18127426e73422f WatchSource:0}: Error finding container 3bffb91d0bd950071f058ad2d50254e43e9cab2b2dc8f73ae18127426e73422f: Status 404 returned error can't find the container with id 3bffb91d0bd950071f058ad2d50254e43e9cab2b2dc8f73ae18127426e73422f Dec 06 03:50:17 crc kubenswrapper[4980]: E1206 03:50:17.738829 4980 upgradeaware.go:427] Error proxying data from client to backend: readfrom tcp 38.102.83.184:39624->38.102.83.184:43319: write tcp 38.102.83.184:39624->38.102.83.184:43319: write: broken pipe Dec 06 03:50:17 crc kubenswrapper[4980]: I1206 03:50:17.964577 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/memcached-0" event={"ID":"cbbe1b5d-848b-4c0f-bd83-10b27906dbfc","Type":"ContainerStarted","Data":"3bffb91d0bd950071f058ad2d50254e43e9cab2b2dc8f73ae18127426e73422f"} Dec 06 03:50:22 crc kubenswrapper[4980]: I1206 03:50:22.173720 4980 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/openstack-galera-0" Dec 06 03:50:22 crc kubenswrapper[4980]: I1206 03:50:22.174258 4980 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/openstack-galera-0" Dec 06 03:50:22 crc kubenswrapper[4980]: I1206 03:50:22.218614 4980 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/openstack-galera-1" Dec 06 03:50:22 crc kubenswrapper[4980]: I1206 03:50:22.225573 4980 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/openstack-galera-1" Dec 06 03:50:22 crc kubenswrapper[4980]: I1206 03:50:22.225635 4980 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/openstack-galera-2" Dec 06 03:50:22 crc kubenswrapper[4980]: I1206 03:50:22.225651 4980 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/openstack-galera-2" Dec 06 03:50:22 crc kubenswrapper[4980]: I1206 03:50:22.291977 4980 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/openstack-galera-2" Dec 06 03:50:23 crc kubenswrapper[4980]: I1206 03:50:23.141422 4980 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/openstack-galera-2" Dec 06 03:50:23 crc kubenswrapper[4980]: E1206 03:50:23.453500 4980 upgradeaware.go:427] Error proxying data from client to backend: readfrom tcp 38.102.83.184:51252->38.102.83.184:43319: write tcp 38.102.83.184:51252->38.102.83.184:43319: write: broken pipe Dec 06 03:50:24 crc kubenswrapper[4980]: I1206 03:50:24.388458 4980 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/rabbitmq-cluster-operator-index-7m77w" Dec 06 03:50:24 crc kubenswrapper[4980]: I1206 03:50:24.389001 4980 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/rabbitmq-cluster-operator-index-7m77w" Dec 06 03:50:24 crc kubenswrapper[4980]: I1206 03:50:24.417360 4980 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/rabbitmq-cluster-operator-index-7m77w" Dec 06 03:50:25 crc kubenswrapper[4980]: I1206 03:50:25.034466 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/memcached-0" event={"ID":"cbbe1b5d-848b-4c0f-bd83-10b27906dbfc","Type":"ContainerStarted","Data":"9220ba51e28495edb50acc574b6c297c1b7bee4108f25a586f9d3818fe1c54b9"} Dec 06 03:50:25 crc kubenswrapper[4980]: I1206 03:50:25.063393 4980 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/memcached-0" podStartSLOduration=2.209505476 podStartE2EDuration="9.063338928s" podCreationTimestamp="2025-12-06 03:50:16 +0000 UTC" firstStartedPulling="2025-12-06 03:50:17.232605467 +0000 UTC m=+976.477356728" lastFinishedPulling="2025-12-06 03:50:24.086438909 +0000 UTC m=+983.331190180" observedRunningTime="2025-12-06 03:50:25.055452261 +0000 UTC m=+984.300203552" watchObservedRunningTime="2025-12-06 03:50:25.063338928 +0000 UTC m=+984.308090209" Dec 06 03:50:25 crc kubenswrapper[4980]: I1206 03:50:25.068921 4980 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/rabbitmq-cluster-operator-index-7m77w" Dec 06 03:50:25 crc kubenswrapper[4980]: I1206 03:50:25.137142 4980 patch_prober.go:28] interesting pod/machine-config-daemon-r5zfb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 03:50:25 crc kubenswrapper[4980]: I1206 03:50:25.137232 4980 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-r5zfb" podUID="320f44d1-a671-4a91-b328-a8b0fdd8f23a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 03:50:25 crc kubenswrapper[4980]: I1206 03:50:25.137298 4980 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-r5zfb" Dec 06 03:50:25 crc kubenswrapper[4980]: I1206 03:50:25.138145 4980 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"89e5bc79489f00d58aeba62f0124b8200c624808c64c4cf50ac8bd11cba63d00"} pod="openshift-machine-config-operator/machine-config-daemon-r5zfb" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 06 03:50:25 crc kubenswrapper[4980]: I1206 03:50:25.138271 4980 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-r5zfb" podUID="320f44d1-a671-4a91-b328-a8b0fdd8f23a" containerName="machine-config-daemon" containerID="cri-o://89e5bc79489f00d58aeba62f0124b8200c624808c64c4cf50ac8bd11cba63d00" gracePeriod=600 Dec 06 03:50:26 crc kubenswrapper[4980]: I1206 03:50:26.042915 4980 generic.go:334] "Generic (PLEG): container finished" podID="320f44d1-a671-4a91-b328-a8b0fdd8f23a" containerID="89e5bc79489f00d58aeba62f0124b8200c624808c64c4cf50ac8bd11cba63d00" exitCode=0 Dec 06 03:50:26 crc kubenswrapper[4980]: I1206 03:50:26.042979 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-r5zfb" event={"ID":"320f44d1-a671-4a91-b328-a8b0fdd8f23a","Type":"ContainerDied","Data":"89e5bc79489f00d58aeba62f0124b8200c624808c64c4cf50ac8bd11cba63d00"} Dec 06 03:50:26 crc kubenswrapper[4980]: I1206 03:50:26.043423 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-r5zfb" event={"ID":"320f44d1-a671-4a91-b328-a8b0fdd8f23a","Type":"ContainerStarted","Data":"a9db17110188a0ff97c6851d4bf74c82d9f9edeef9b7cfcb0fd15528f7d2e059"} Dec 06 03:50:26 crc kubenswrapper[4980]: I1206 03:50:26.043463 4980 scope.go:117] "RemoveContainer" containerID="c1b55d544de848b8b12ac6ab1d4d7ce58e20980d22589239fd7500d5b466d62e" Dec 06 03:50:26 crc kubenswrapper[4980]: I1206 03:50:26.043589 4980 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/memcached-0" Dec 06 03:50:27 crc kubenswrapper[4980]: I1206 03:50:27.063413 4980 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590cnjz7"] Dec 06 03:50:27 crc kubenswrapper[4980]: I1206 03:50:27.064899 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590cnjz7" Dec 06 03:50:27 crc kubenswrapper[4980]: I1206 03:50:27.066701 4980 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"default-dockercfg-brs8h" Dec 06 03:50:27 crc kubenswrapper[4980]: I1206 03:50:27.076569 4980 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590cnjz7"] Dec 06 03:50:27 crc kubenswrapper[4980]: I1206 03:50:27.100758 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/b2cb8251-b0fb-4d95-a2fb-460ab17081e4-bundle\") pod \"9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590cnjz7\" (UID: \"b2cb8251-b0fb-4d95-a2fb-460ab17081e4\") " pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590cnjz7" Dec 06 03:50:27 crc kubenswrapper[4980]: I1206 03:50:27.100830 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/b2cb8251-b0fb-4d95-a2fb-460ab17081e4-util\") pod \"9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590cnjz7\" (UID: \"b2cb8251-b0fb-4d95-a2fb-460ab17081e4\") " pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590cnjz7" Dec 06 03:50:27 crc kubenswrapper[4980]: I1206 03:50:27.100925 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tf8q9\" (UniqueName: \"kubernetes.io/projected/b2cb8251-b0fb-4d95-a2fb-460ab17081e4-kube-api-access-tf8q9\") pod \"9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590cnjz7\" (UID: \"b2cb8251-b0fb-4d95-a2fb-460ab17081e4\") " pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590cnjz7" Dec 06 03:50:27 crc kubenswrapper[4980]: I1206 03:50:27.201951 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tf8q9\" (UniqueName: \"kubernetes.io/projected/b2cb8251-b0fb-4d95-a2fb-460ab17081e4-kube-api-access-tf8q9\") pod \"9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590cnjz7\" (UID: \"b2cb8251-b0fb-4d95-a2fb-460ab17081e4\") " pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590cnjz7" Dec 06 03:50:27 crc kubenswrapper[4980]: I1206 03:50:27.202025 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/b2cb8251-b0fb-4d95-a2fb-460ab17081e4-bundle\") pod \"9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590cnjz7\" (UID: \"b2cb8251-b0fb-4d95-a2fb-460ab17081e4\") " pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590cnjz7" Dec 06 03:50:27 crc kubenswrapper[4980]: I1206 03:50:27.202062 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/b2cb8251-b0fb-4d95-a2fb-460ab17081e4-util\") pod \"9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590cnjz7\" (UID: \"b2cb8251-b0fb-4d95-a2fb-460ab17081e4\") " pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590cnjz7" Dec 06 03:50:27 crc kubenswrapper[4980]: I1206 03:50:27.202629 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/b2cb8251-b0fb-4d95-a2fb-460ab17081e4-bundle\") pod \"9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590cnjz7\" (UID: \"b2cb8251-b0fb-4d95-a2fb-460ab17081e4\") " pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590cnjz7" Dec 06 03:50:27 crc kubenswrapper[4980]: I1206 03:50:27.202649 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/b2cb8251-b0fb-4d95-a2fb-460ab17081e4-util\") pod \"9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590cnjz7\" (UID: \"b2cb8251-b0fb-4d95-a2fb-460ab17081e4\") " pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590cnjz7" Dec 06 03:50:27 crc kubenswrapper[4980]: I1206 03:50:27.238997 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tf8q9\" (UniqueName: \"kubernetes.io/projected/b2cb8251-b0fb-4d95-a2fb-460ab17081e4-kube-api-access-tf8q9\") pod \"9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590cnjz7\" (UID: \"b2cb8251-b0fb-4d95-a2fb-460ab17081e4\") " pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590cnjz7" Dec 06 03:50:27 crc kubenswrapper[4980]: I1206 03:50:27.401457 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590cnjz7" Dec 06 03:50:29 crc kubenswrapper[4980]: I1206 03:50:29.000735 4980 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590cnjz7"] Dec 06 03:50:29 crc kubenswrapper[4980]: W1206 03:50:29.009946 4980 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb2cb8251_b0fb_4d95_a2fb_460ab17081e4.slice/crio-031e7ca7e7791e21892c5634211963501681c3940087fc26de366d464be842bb WatchSource:0}: Error finding container 031e7ca7e7791e21892c5634211963501681c3940087fc26de366d464be842bb: Status 404 returned error can't find the container with id 031e7ca7e7791e21892c5634211963501681c3940087fc26de366d464be842bb Dec 06 03:50:29 crc kubenswrapper[4980]: I1206 03:50:29.063816 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590cnjz7" event={"ID":"b2cb8251-b0fb-4d95-a2fb-460ab17081e4","Type":"ContainerStarted","Data":"031e7ca7e7791e21892c5634211963501681c3940087fc26de366d464be842bb"} Dec 06 03:50:30 crc kubenswrapper[4980]: I1206 03:50:30.072188 4980 generic.go:334] "Generic (PLEG): container finished" podID="b2cb8251-b0fb-4d95-a2fb-460ab17081e4" containerID="1d64fb1b1645438df6f7dd658a1e407bde3361eeb6aac2cd419fbbf9fb095b32" exitCode=0 Dec 06 03:50:30 crc kubenswrapper[4980]: I1206 03:50:30.072308 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590cnjz7" event={"ID":"b2cb8251-b0fb-4d95-a2fb-460ab17081e4","Type":"ContainerDied","Data":"1d64fb1b1645438df6f7dd658a1e407bde3361eeb6aac2cd419fbbf9fb095b32"} Dec 06 03:50:31 crc kubenswrapper[4980]: I1206 03:50:31.231308 4980 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-bqts9"] Dec 06 03:50:31 crc kubenswrapper[4980]: I1206 03:50:31.233469 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-bqts9" Dec 06 03:50:31 crc kubenswrapper[4980]: I1206 03:50:31.246847 4980 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-bqts9"] Dec 06 03:50:31 crc kubenswrapper[4980]: I1206 03:50:31.365471 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jv8hf\" (UniqueName: \"kubernetes.io/projected/60462c94-2c21-42ce-9cdf-b4fb988cabaa-kube-api-access-jv8hf\") pod \"community-operators-bqts9\" (UID: \"60462c94-2c21-42ce-9cdf-b4fb988cabaa\") " pod="openshift-marketplace/community-operators-bqts9" Dec 06 03:50:31 crc kubenswrapper[4980]: I1206 03:50:31.365586 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/60462c94-2c21-42ce-9cdf-b4fb988cabaa-utilities\") pod \"community-operators-bqts9\" (UID: \"60462c94-2c21-42ce-9cdf-b4fb988cabaa\") " pod="openshift-marketplace/community-operators-bqts9" Dec 06 03:50:31 crc kubenswrapper[4980]: I1206 03:50:31.365605 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/60462c94-2c21-42ce-9cdf-b4fb988cabaa-catalog-content\") pod \"community-operators-bqts9\" (UID: \"60462c94-2c21-42ce-9cdf-b4fb988cabaa\") " pod="openshift-marketplace/community-operators-bqts9" Dec 06 03:50:31 crc kubenswrapper[4980]: I1206 03:50:31.466899 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jv8hf\" (UniqueName: \"kubernetes.io/projected/60462c94-2c21-42ce-9cdf-b4fb988cabaa-kube-api-access-jv8hf\") pod \"community-operators-bqts9\" (UID: \"60462c94-2c21-42ce-9cdf-b4fb988cabaa\") " pod="openshift-marketplace/community-operators-bqts9" Dec 06 03:50:31 crc kubenswrapper[4980]: I1206 03:50:31.467008 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/60462c94-2c21-42ce-9cdf-b4fb988cabaa-utilities\") pod \"community-operators-bqts9\" (UID: \"60462c94-2c21-42ce-9cdf-b4fb988cabaa\") " pod="openshift-marketplace/community-operators-bqts9" Dec 06 03:50:31 crc kubenswrapper[4980]: I1206 03:50:31.467038 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/60462c94-2c21-42ce-9cdf-b4fb988cabaa-catalog-content\") pod \"community-operators-bqts9\" (UID: \"60462c94-2c21-42ce-9cdf-b4fb988cabaa\") " pod="openshift-marketplace/community-operators-bqts9" Dec 06 03:50:31 crc kubenswrapper[4980]: I1206 03:50:31.467621 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/60462c94-2c21-42ce-9cdf-b4fb988cabaa-utilities\") pod \"community-operators-bqts9\" (UID: \"60462c94-2c21-42ce-9cdf-b4fb988cabaa\") " pod="openshift-marketplace/community-operators-bqts9" Dec 06 03:50:31 crc kubenswrapper[4980]: I1206 03:50:31.467686 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/60462c94-2c21-42ce-9cdf-b4fb988cabaa-catalog-content\") pod \"community-operators-bqts9\" (UID: \"60462c94-2c21-42ce-9cdf-b4fb988cabaa\") " pod="openshift-marketplace/community-operators-bqts9" Dec 06 03:50:31 crc kubenswrapper[4980]: I1206 03:50:31.497435 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jv8hf\" (UniqueName: \"kubernetes.io/projected/60462c94-2c21-42ce-9cdf-b4fb988cabaa-kube-api-access-jv8hf\") pod \"community-operators-bqts9\" (UID: \"60462c94-2c21-42ce-9cdf-b4fb988cabaa\") " pod="openshift-marketplace/community-operators-bqts9" Dec 06 03:50:31 crc kubenswrapper[4980]: I1206 03:50:31.553430 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-bqts9" Dec 06 03:50:31 crc kubenswrapper[4980]: I1206 03:50:31.819752 4980 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/memcached-0" Dec 06 03:50:32 crc kubenswrapper[4980]: I1206 03:50:32.093236 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590cnjz7" event={"ID":"b2cb8251-b0fb-4d95-a2fb-460ab17081e4","Type":"ContainerStarted","Data":"dc35755469fc64fde642bf8013cd13f266b137e510588ddc7c6a2e999156f66f"} Dec 06 03:50:32 crc kubenswrapper[4980]: I1206 03:50:32.120095 4980 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-bqts9"] Dec 06 03:50:32 crc kubenswrapper[4980]: W1206 03:50:32.188627 4980 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod60462c94_2c21_42ce_9cdf_b4fb988cabaa.slice/crio-e71b8cc0b19f244fedb4ea25c29136a4c5290146093c7a4ebe8b83f327e322d5 WatchSource:0}: Error finding container e71b8cc0b19f244fedb4ea25c29136a4c5290146093c7a4ebe8b83f327e322d5: Status 404 returned error can't find the container with id e71b8cc0b19f244fedb4ea25c29136a4c5290146093c7a4ebe8b83f327e322d5 Dec 06 03:50:32 crc kubenswrapper[4980]: I1206 03:50:32.262084 4980 prober.go:107] "Probe failed" probeType="Readiness" pod="glance-kuttl-tests/openstack-galera-2" podUID="c53b1fad-da29-4e9c-970c-9f4d84693905" containerName="galera" probeResult="failure" output=< Dec 06 03:50:32 crc kubenswrapper[4980]: wsrep_local_state_comment (Donor/Desynced) differs from Synced Dec 06 03:50:32 crc kubenswrapper[4980]: > Dec 06 03:50:33 crc kubenswrapper[4980]: I1206 03:50:33.125012 4980 generic.go:334] "Generic (PLEG): container finished" podID="60462c94-2c21-42ce-9cdf-b4fb988cabaa" containerID="88e0e7746c9cd6839fbfaaffbf6b359b1a8e47d9c86e7b0094e22381419333f7" exitCode=0 Dec 06 03:50:33 crc kubenswrapper[4980]: I1206 03:50:33.125445 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-bqts9" event={"ID":"60462c94-2c21-42ce-9cdf-b4fb988cabaa","Type":"ContainerDied","Data":"88e0e7746c9cd6839fbfaaffbf6b359b1a8e47d9c86e7b0094e22381419333f7"} Dec 06 03:50:33 crc kubenswrapper[4980]: I1206 03:50:33.125500 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-bqts9" event={"ID":"60462c94-2c21-42ce-9cdf-b4fb988cabaa","Type":"ContainerStarted","Data":"e71b8cc0b19f244fedb4ea25c29136a4c5290146093c7a4ebe8b83f327e322d5"} Dec 06 03:50:33 crc kubenswrapper[4980]: I1206 03:50:33.138606 4980 generic.go:334] "Generic (PLEG): container finished" podID="b2cb8251-b0fb-4d95-a2fb-460ab17081e4" containerID="dc35755469fc64fde642bf8013cd13f266b137e510588ddc7c6a2e999156f66f" exitCode=0 Dec 06 03:50:33 crc kubenswrapper[4980]: I1206 03:50:33.138676 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590cnjz7" event={"ID":"b2cb8251-b0fb-4d95-a2fb-460ab17081e4","Type":"ContainerDied","Data":"dc35755469fc64fde642bf8013cd13f266b137e510588ddc7c6a2e999156f66f"} Dec 06 03:50:34 crc kubenswrapper[4980]: I1206 03:50:34.149610 4980 generic.go:334] "Generic (PLEG): container finished" podID="b2cb8251-b0fb-4d95-a2fb-460ab17081e4" containerID="e8fb98ce798b3475b7914b29031eb091fc026ca9a623b78cdd22c71b025d03e6" exitCode=0 Dec 06 03:50:34 crc kubenswrapper[4980]: I1206 03:50:34.149907 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590cnjz7" event={"ID":"b2cb8251-b0fb-4d95-a2fb-460ab17081e4","Type":"ContainerDied","Data":"e8fb98ce798b3475b7914b29031eb091fc026ca9a623b78cdd22c71b025d03e6"} Dec 06 03:50:34 crc kubenswrapper[4980]: I1206 03:50:34.163952 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-bqts9" event={"ID":"60462c94-2c21-42ce-9cdf-b4fb988cabaa","Type":"ContainerStarted","Data":"50ce4df92ebca16e3a33ae4f66d8ff24dead139794a4cd29b2019053cf438c5e"} Dec 06 03:50:35 crc kubenswrapper[4980]: I1206 03:50:35.173000 4980 generic.go:334] "Generic (PLEG): container finished" podID="60462c94-2c21-42ce-9cdf-b4fb988cabaa" containerID="50ce4df92ebca16e3a33ae4f66d8ff24dead139794a4cd29b2019053cf438c5e" exitCode=0 Dec 06 03:50:35 crc kubenswrapper[4980]: I1206 03:50:35.173043 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-bqts9" event={"ID":"60462c94-2c21-42ce-9cdf-b4fb988cabaa","Type":"ContainerDied","Data":"50ce4df92ebca16e3a33ae4f66d8ff24dead139794a4cd29b2019053cf438c5e"} Dec 06 03:50:35 crc kubenswrapper[4980]: I1206 03:50:35.478966 4980 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590cnjz7" Dec 06 03:50:35 crc kubenswrapper[4980]: I1206 03:50:35.524280 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/b2cb8251-b0fb-4d95-a2fb-460ab17081e4-util\") pod \"b2cb8251-b0fb-4d95-a2fb-460ab17081e4\" (UID: \"b2cb8251-b0fb-4d95-a2fb-460ab17081e4\") " Dec 06 03:50:35 crc kubenswrapper[4980]: I1206 03:50:35.524403 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/b2cb8251-b0fb-4d95-a2fb-460ab17081e4-bundle\") pod \"b2cb8251-b0fb-4d95-a2fb-460ab17081e4\" (UID: \"b2cb8251-b0fb-4d95-a2fb-460ab17081e4\") " Dec 06 03:50:35 crc kubenswrapper[4980]: I1206 03:50:35.524460 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tf8q9\" (UniqueName: \"kubernetes.io/projected/b2cb8251-b0fb-4d95-a2fb-460ab17081e4-kube-api-access-tf8q9\") pod \"b2cb8251-b0fb-4d95-a2fb-460ab17081e4\" (UID: \"b2cb8251-b0fb-4d95-a2fb-460ab17081e4\") " Dec 06 03:50:35 crc kubenswrapper[4980]: I1206 03:50:35.525052 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b2cb8251-b0fb-4d95-a2fb-460ab17081e4-bundle" (OuterVolumeSpecName: "bundle") pod "b2cb8251-b0fb-4d95-a2fb-460ab17081e4" (UID: "b2cb8251-b0fb-4d95-a2fb-460ab17081e4"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 03:50:35 crc kubenswrapper[4980]: I1206 03:50:35.536890 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b2cb8251-b0fb-4d95-a2fb-460ab17081e4-util" (OuterVolumeSpecName: "util") pod "b2cb8251-b0fb-4d95-a2fb-460ab17081e4" (UID: "b2cb8251-b0fb-4d95-a2fb-460ab17081e4"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 03:50:35 crc kubenswrapper[4980]: I1206 03:50:35.550107 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b2cb8251-b0fb-4d95-a2fb-460ab17081e4-kube-api-access-tf8q9" (OuterVolumeSpecName: "kube-api-access-tf8q9") pod "b2cb8251-b0fb-4d95-a2fb-460ab17081e4" (UID: "b2cb8251-b0fb-4d95-a2fb-460ab17081e4"). InnerVolumeSpecName "kube-api-access-tf8q9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 03:50:35 crc kubenswrapper[4980]: I1206 03:50:35.626349 4980 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/b2cb8251-b0fb-4d95-a2fb-460ab17081e4-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 03:50:35 crc kubenswrapper[4980]: I1206 03:50:35.626383 4980 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tf8q9\" (UniqueName: \"kubernetes.io/projected/b2cb8251-b0fb-4d95-a2fb-460ab17081e4-kube-api-access-tf8q9\") on node \"crc\" DevicePath \"\"" Dec 06 03:50:35 crc kubenswrapper[4980]: I1206 03:50:35.626396 4980 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/b2cb8251-b0fb-4d95-a2fb-460ab17081e4-util\") on node \"crc\" DevicePath \"\"" Dec 06 03:50:36 crc kubenswrapper[4980]: I1206 03:50:36.183524 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-bqts9" event={"ID":"60462c94-2c21-42ce-9cdf-b4fb988cabaa","Type":"ContainerStarted","Data":"f964025a1b99d2453d75bd3998738ea81e89cf19e25b48a164862aaf755d479a"} Dec 06 03:50:36 crc kubenswrapper[4980]: I1206 03:50:36.186847 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590cnjz7" event={"ID":"b2cb8251-b0fb-4d95-a2fb-460ab17081e4","Type":"ContainerDied","Data":"031e7ca7e7791e21892c5634211963501681c3940087fc26de366d464be842bb"} Dec 06 03:50:36 crc kubenswrapper[4980]: I1206 03:50:36.186881 4980 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="031e7ca7e7791e21892c5634211963501681c3940087fc26de366d464be842bb" Dec 06 03:50:36 crc kubenswrapper[4980]: I1206 03:50:36.186947 4980 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590cnjz7" Dec 06 03:50:36 crc kubenswrapper[4980]: I1206 03:50:36.205320 4980 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-bqts9" podStartSLOduration=2.650487 podStartE2EDuration="5.205286614s" podCreationTimestamp="2025-12-06 03:50:31 +0000 UTC" firstStartedPulling="2025-12-06 03:50:33.134340108 +0000 UTC m=+992.379091379" lastFinishedPulling="2025-12-06 03:50:35.689139732 +0000 UTC m=+994.933890993" observedRunningTime="2025-12-06 03:50:36.201774203 +0000 UTC m=+995.446525494" watchObservedRunningTime="2025-12-06 03:50:36.205286614 +0000 UTC m=+995.450037885" Dec 06 03:50:36 crc kubenswrapper[4980]: I1206 03:50:36.646529 4980 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/openstack-galera-0" Dec 06 03:50:36 crc kubenswrapper[4980]: I1206 03:50:36.699072 4980 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/openstack-galera-0" Dec 06 03:50:36 crc kubenswrapper[4980]: I1206 03:50:36.701620 4980 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/openstack-galera-1" Dec 06 03:50:36 crc kubenswrapper[4980]: I1206 03:50:36.881353 4980 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/openstack-galera-1" Dec 06 03:50:41 crc kubenswrapper[4980]: I1206 03:50:41.554079 4980 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-bqts9" Dec 06 03:50:41 crc kubenswrapper[4980]: I1206 03:50:41.554684 4980 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-bqts9" Dec 06 03:50:41 crc kubenswrapper[4980]: I1206 03:50:41.596121 4980 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-bqts9" Dec 06 03:50:42 crc kubenswrapper[4980]: I1206 03:50:42.267106 4980 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-bqts9" Dec 06 03:50:44 crc kubenswrapper[4980]: I1206 03:50:44.670545 4980 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-779fc9694b-r7tt8"] Dec 06 03:50:44 crc kubenswrapper[4980]: E1206 03:50:44.671071 4980 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b2cb8251-b0fb-4d95-a2fb-460ab17081e4" containerName="pull" Dec 06 03:50:44 crc kubenswrapper[4980]: I1206 03:50:44.671086 4980 state_mem.go:107] "Deleted CPUSet assignment" podUID="b2cb8251-b0fb-4d95-a2fb-460ab17081e4" containerName="pull" Dec 06 03:50:44 crc kubenswrapper[4980]: E1206 03:50:44.671101 4980 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b2cb8251-b0fb-4d95-a2fb-460ab17081e4" containerName="util" Dec 06 03:50:44 crc kubenswrapper[4980]: I1206 03:50:44.671109 4980 state_mem.go:107] "Deleted CPUSet assignment" podUID="b2cb8251-b0fb-4d95-a2fb-460ab17081e4" containerName="util" Dec 06 03:50:44 crc kubenswrapper[4980]: E1206 03:50:44.671118 4980 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b2cb8251-b0fb-4d95-a2fb-460ab17081e4" containerName="extract" Dec 06 03:50:44 crc kubenswrapper[4980]: I1206 03:50:44.671127 4980 state_mem.go:107] "Deleted CPUSet assignment" podUID="b2cb8251-b0fb-4d95-a2fb-460ab17081e4" containerName="extract" Dec 06 03:50:44 crc kubenswrapper[4980]: I1206 03:50:44.671262 4980 memory_manager.go:354] "RemoveStaleState removing state" podUID="b2cb8251-b0fb-4d95-a2fb-460ab17081e4" containerName="extract" Dec 06 03:50:44 crc kubenswrapper[4980]: I1206 03:50:44.671858 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-779fc9694b-r7tt8" Dec 06 03:50:44 crc kubenswrapper[4980]: I1206 03:50:44.674918 4980 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"rabbitmq-cluster-operator-dockercfg-tfsj7" Dec 06 03:50:44 crc kubenswrapper[4980]: I1206 03:50:44.695144 4980 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-779fc9694b-r7tt8"] Dec 06 03:50:44 crc kubenswrapper[4980]: I1206 03:50:44.697902 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bnqf8\" (UniqueName: \"kubernetes.io/projected/003f4f15-503c-4321-9daa-46f4ca2541d1-kube-api-access-bnqf8\") pod \"rabbitmq-cluster-operator-779fc9694b-r7tt8\" (UID: \"003f4f15-503c-4321-9daa-46f4ca2541d1\") " pod="openstack-operators/rabbitmq-cluster-operator-779fc9694b-r7tt8" Dec 06 03:50:44 crc kubenswrapper[4980]: I1206 03:50:44.799481 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bnqf8\" (UniqueName: \"kubernetes.io/projected/003f4f15-503c-4321-9daa-46f4ca2541d1-kube-api-access-bnqf8\") pod \"rabbitmq-cluster-operator-779fc9694b-r7tt8\" (UID: \"003f4f15-503c-4321-9daa-46f4ca2541d1\") " pod="openstack-operators/rabbitmq-cluster-operator-779fc9694b-r7tt8" Dec 06 03:50:44 crc kubenswrapper[4980]: I1206 03:50:44.835325 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bnqf8\" (UniqueName: \"kubernetes.io/projected/003f4f15-503c-4321-9daa-46f4ca2541d1-kube-api-access-bnqf8\") pod \"rabbitmq-cluster-operator-779fc9694b-r7tt8\" (UID: \"003f4f15-503c-4321-9daa-46f4ca2541d1\") " pod="openstack-operators/rabbitmq-cluster-operator-779fc9694b-r7tt8" Dec 06 03:50:45 crc kubenswrapper[4980]: I1206 03:50:44.999923 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-779fc9694b-r7tt8" Dec 06 03:50:45 crc kubenswrapper[4980]: I1206 03:50:45.021350 4980 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-bqts9"] Dec 06 03:50:45 crc kubenswrapper[4980]: I1206 03:50:45.021639 4980 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-bqts9" podUID="60462c94-2c21-42ce-9cdf-b4fb988cabaa" containerName="registry-server" containerID="cri-o://f964025a1b99d2453d75bd3998738ea81e89cf19e25b48a164862aaf755d479a" gracePeriod=2 Dec 06 03:50:45 crc kubenswrapper[4980]: I1206 03:50:45.422936 4980 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-779fc9694b-r7tt8"] Dec 06 03:50:46 crc kubenswrapper[4980]: I1206 03:50:46.568395 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-779fc9694b-r7tt8" event={"ID":"003f4f15-503c-4321-9daa-46f4ca2541d1","Type":"ContainerStarted","Data":"ed128b6c134ff792335aee04432a91e12060979c077a64099a2c896f2f4f99b0"} Dec 06 03:50:46 crc kubenswrapper[4980]: I1206 03:50:46.579913 4980 generic.go:334] "Generic (PLEG): container finished" podID="60462c94-2c21-42ce-9cdf-b4fb988cabaa" containerID="f964025a1b99d2453d75bd3998738ea81e89cf19e25b48a164862aaf755d479a" exitCode=0 Dec 06 03:50:46 crc kubenswrapper[4980]: I1206 03:50:46.580024 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-bqts9" event={"ID":"60462c94-2c21-42ce-9cdf-b4fb988cabaa","Type":"ContainerDied","Data":"f964025a1b99d2453d75bd3998738ea81e89cf19e25b48a164862aaf755d479a"} Dec 06 03:50:46 crc kubenswrapper[4980]: I1206 03:50:46.927871 4980 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-bqts9" Dec 06 03:50:47 crc kubenswrapper[4980]: I1206 03:50:47.028277 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/60462c94-2c21-42ce-9cdf-b4fb988cabaa-utilities\") pod \"60462c94-2c21-42ce-9cdf-b4fb988cabaa\" (UID: \"60462c94-2c21-42ce-9cdf-b4fb988cabaa\") " Dec 06 03:50:47 crc kubenswrapper[4980]: I1206 03:50:47.029658 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/60462c94-2c21-42ce-9cdf-b4fb988cabaa-utilities" (OuterVolumeSpecName: "utilities") pod "60462c94-2c21-42ce-9cdf-b4fb988cabaa" (UID: "60462c94-2c21-42ce-9cdf-b4fb988cabaa"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 03:50:47 crc kubenswrapper[4980]: I1206 03:50:47.130127 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/60462c94-2c21-42ce-9cdf-b4fb988cabaa-catalog-content\") pod \"60462c94-2c21-42ce-9cdf-b4fb988cabaa\" (UID: \"60462c94-2c21-42ce-9cdf-b4fb988cabaa\") " Dec 06 03:50:47 crc kubenswrapper[4980]: I1206 03:50:47.130226 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jv8hf\" (UniqueName: \"kubernetes.io/projected/60462c94-2c21-42ce-9cdf-b4fb988cabaa-kube-api-access-jv8hf\") pod \"60462c94-2c21-42ce-9cdf-b4fb988cabaa\" (UID: \"60462c94-2c21-42ce-9cdf-b4fb988cabaa\") " Dec 06 03:50:47 crc kubenswrapper[4980]: I1206 03:50:47.130441 4980 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/60462c94-2c21-42ce-9cdf-b4fb988cabaa-utilities\") on node \"crc\" DevicePath \"\"" Dec 06 03:50:47 crc kubenswrapper[4980]: I1206 03:50:47.137754 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/60462c94-2c21-42ce-9cdf-b4fb988cabaa-kube-api-access-jv8hf" (OuterVolumeSpecName: "kube-api-access-jv8hf") pod "60462c94-2c21-42ce-9cdf-b4fb988cabaa" (UID: "60462c94-2c21-42ce-9cdf-b4fb988cabaa"). InnerVolumeSpecName "kube-api-access-jv8hf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 03:50:47 crc kubenswrapper[4980]: I1206 03:50:47.177638 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/60462c94-2c21-42ce-9cdf-b4fb988cabaa-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "60462c94-2c21-42ce-9cdf-b4fb988cabaa" (UID: "60462c94-2c21-42ce-9cdf-b4fb988cabaa"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 03:50:47 crc kubenswrapper[4980]: I1206 03:50:47.231458 4980 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/60462c94-2c21-42ce-9cdf-b4fb988cabaa-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 06 03:50:47 crc kubenswrapper[4980]: I1206 03:50:47.231528 4980 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jv8hf\" (UniqueName: \"kubernetes.io/projected/60462c94-2c21-42ce-9cdf-b4fb988cabaa-kube-api-access-jv8hf\") on node \"crc\" DevicePath \"\"" Dec 06 03:50:47 crc kubenswrapper[4980]: I1206 03:50:47.588677 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-bqts9" event={"ID":"60462c94-2c21-42ce-9cdf-b4fb988cabaa","Type":"ContainerDied","Data":"e71b8cc0b19f244fedb4ea25c29136a4c5290146093c7a4ebe8b83f327e322d5"} Dec 06 03:50:47 crc kubenswrapper[4980]: I1206 03:50:47.588777 4980 scope.go:117] "RemoveContainer" containerID="f964025a1b99d2453d75bd3998738ea81e89cf19e25b48a164862aaf755d479a" Dec 06 03:50:47 crc kubenswrapper[4980]: I1206 03:50:47.588786 4980 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-bqts9" Dec 06 03:50:47 crc kubenswrapper[4980]: I1206 03:50:47.616398 4980 scope.go:117] "RemoveContainer" containerID="50ce4df92ebca16e3a33ae4f66d8ff24dead139794a4cd29b2019053cf438c5e" Dec 06 03:50:47 crc kubenswrapper[4980]: I1206 03:50:47.619631 4980 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-bqts9"] Dec 06 03:50:47 crc kubenswrapper[4980]: I1206 03:50:47.620339 4980 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-bqts9"] Dec 06 03:50:47 crc kubenswrapper[4980]: I1206 03:50:47.644272 4980 scope.go:117] "RemoveContainer" containerID="88e0e7746c9cd6839fbfaaffbf6b359b1a8e47d9c86e7b0094e22381419333f7" Dec 06 03:50:49 crc kubenswrapper[4980]: I1206 03:50:49.195247 4980 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="60462c94-2c21-42ce-9cdf-b4fb988cabaa" path="/var/lib/kubelet/pods/60462c94-2c21-42ce-9cdf-b4fb988cabaa/volumes" Dec 06 03:50:50 crc kubenswrapper[4980]: I1206 03:50:50.614286 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-779fc9694b-r7tt8" event={"ID":"003f4f15-503c-4321-9daa-46f4ca2541d1","Type":"ContainerStarted","Data":"2be5d874e9544cf5e49c1137981d59b0f013dc19809e9bf3bba4e49bd7e37c20"} Dec 06 03:50:50 crc kubenswrapper[4980]: I1206 03:50:50.639573 4980 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/rabbitmq-cluster-operator-779fc9694b-r7tt8" podStartSLOduration=2.6084852080000003 podStartE2EDuration="6.639505129s" podCreationTimestamp="2025-12-06 03:50:44 +0000 UTC" firstStartedPulling="2025-12-06 03:50:45.473533648 +0000 UTC m=+1004.718284929" lastFinishedPulling="2025-12-06 03:50:49.504553579 +0000 UTC m=+1008.749304850" observedRunningTime="2025-12-06 03:50:50.6381396 +0000 UTC m=+1009.882890881" watchObservedRunningTime="2025-12-06 03:50:50.639505129 +0000 UTC m=+1009.884256420" Dec 06 03:50:51 crc kubenswrapper[4980]: I1206 03:50:51.234418 4980 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-6pbbz"] Dec 06 03:50:51 crc kubenswrapper[4980]: E1206 03:50:51.234701 4980 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="60462c94-2c21-42ce-9cdf-b4fb988cabaa" containerName="extract-content" Dec 06 03:50:51 crc kubenswrapper[4980]: I1206 03:50:51.234718 4980 state_mem.go:107] "Deleted CPUSet assignment" podUID="60462c94-2c21-42ce-9cdf-b4fb988cabaa" containerName="extract-content" Dec 06 03:50:51 crc kubenswrapper[4980]: E1206 03:50:51.234738 4980 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="60462c94-2c21-42ce-9cdf-b4fb988cabaa" containerName="extract-utilities" Dec 06 03:50:51 crc kubenswrapper[4980]: I1206 03:50:51.234746 4980 state_mem.go:107] "Deleted CPUSet assignment" podUID="60462c94-2c21-42ce-9cdf-b4fb988cabaa" containerName="extract-utilities" Dec 06 03:50:51 crc kubenswrapper[4980]: E1206 03:50:51.234755 4980 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="60462c94-2c21-42ce-9cdf-b4fb988cabaa" containerName="registry-server" Dec 06 03:50:51 crc kubenswrapper[4980]: I1206 03:50:51.234761 4980 state_mem.go:107] "Deleted CPUSet assignment" podUID="60462c94-2c21-42ce-9cdf-b4fb988cabaa" containerName="registry-server" Dec 06 03:50:51 crc kubenswrapper[4980]: I1206 03:50:51.234873 4980 memory_manager.go:354] "RemoveStaleState removing state" podUID="60462c94-2c21-42ce-9cdf-b4fb988cabaa" containerName="registry-server" Dec 06 03:50:51 crc kubenswrapper[4980]: I1206 03:50:51.235797 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-6pbbz" Dec 06 03:50:51 crc kubenswrapper[4980]: I1206 03:50:51.274722 4980 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-6pbbz"] Dec 06 03:50:51 crc kubenswrapper[4980]: I1206 03:50:51.398596 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w588x\" (UniqueName: \"kubernetes.io/projected/59a18727-c4b3-488f-ad90-08c32fd14878-kube-api-access-w588x\") pod \"redhat-marketplace-6pbbz\" (UID: \"59a18727-c4b3-488f-ad90-08c32fd14878\") " pod="openshift-marketplace/redhat-marketplace-6pbbz" Dec 06 03:50:51 crc kubenswrapper[4980]: I1206 03:50:51.398681 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/59a18727-c4b3-488f-ad90-08c32fd14878-utilities\") pod \"redhat-marketplace-6pbbz\" (UID: \"59a18727-c4b3-488f-ad90-08c32fd14878\") " pod="openshift-marketplace/redhat-marketplace-6pbbz" Dec 06 03:50:51 crc kubenswrapper[4980]: I1206 03:50:51.398723 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/59a18727-c4b3-488f-ad90-08c32fd14878-catalog-content\") pod \"redhat-marketplace-6pbbz\" (UID: \"59a18727-c4b3-488f-ad90-08c32fd14878\") " pod="openshift-marketplace/redhat-marketplace-6pbbz" Dec 06 03:50:51 crc kubenswrapper[4980]: I1206 03:50:51.500262 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w588x\" (UniqueName: \"kubernetes.io/projected/59a18727-c4b3-488f-ad90-08c32fd14878-kube-api-access-w588x\") pod \"redhat-marketplace-6pbbz\" (UID: \"59a18727-c4b3-488f-ad90-08c32fd14878\") " pod="openshift-marketplace/redhat-marketplace-6pbbz" Dec 06 03:50:51 crc kubenswrapper[4980]: I1206 03:50:51.500348 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/59a18727-c4b3-488f-ad90-08c32fd14878-utilities\") pod \"redhat-marketplace-6pbbz\" (UID: \"59a18727-c4b3-488f-ad90-08c32fd14878\") " pod="openshift-marketplace/redhat-marketplace-6pbbz" Dec 06 03:50:51 crc kubenswrapper[4980]: I1206 03:50:51.500400 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/59a18727-c4b3-488f-ad90-08c32fd14878-catalog-content\") pod \"redhat-marketplace-6pbbz\" (UID: \"59a18727-c4b3-488f-ad90-08c32fd14878\") " pod="openshift-marketplace/redhat-marketplace-6pbbz" Dec 06 03:50:51 crc kubenswrapper[4980]: I1206 03:50:51.501079 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/59a18727-c4b3-488f-ad90-08c32fd14878-utilities\") pod \"redhat-marketplace-6pbbz\" (UID: \"59a18727-c4b3-488f-ad90-08c32fd14878\") " pod="openshift-marketplace/redhat-marketplace-6pbbz" Dec 06 03:50:51 crc kubenswrapper[4980]: I1206 03:50:51.501146 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/59a18727-c4b3-488f-ad90-08c32fd14878-catalog-content\") pod \"redhat-marketplace-6pbbz\" (UID: \"59a18727-c4b3-488f-ad90-08c32fd14878\") " pod="openshift-marketplace/redhat-marketplace-6pbbz" Dec 06 03:50:51 crc kubenswrapper[4980]: I1206 03:50:51.550407 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w588x\" (UniqueName: \"kubernetes.io/projected/59a18727-c4b3-488f-ad90-08c32fd14878-kube-api-access-w588x\") pod \"redhat-marketplace-6pbbz\" (UID: \"59a18727-c4b3-488f-ad90-08c32fd14878\") " pod="openshift-marketplace/redhat-marketplace-6pbbz" Dec 06 03:50:51 crc kubenswrapper[4980]: I1206 03:50:51.553241 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-6pbbz" Dec 06 03:50:52 crc kubenswrapper[4980]: I1206 03:50:52.013195 4980 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-6pbbz"] Dec 06 03:50:52 crc kubenswrapper[4980]: W1206 03:50:52.017770 4980 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod59a18727_c4b3_488f_ad90_08c32fd14878.slice/crio-0e9314dededba865225a172203c5842f1b0d94a7faa663445e20ed522e5b2ed9 WatchSource:0}: Error finding container 0e9314dededba865225a172203c5842f1b0d94a7faa663445e20ed522e5b2ed9: Status 404 returned error can't find the container with id 0e9314dededba865225a172203c5842f1b0d94a7faa663445e20ed522e5b2ed9 Dec 06 03:50:52 crc kubenswrapper[4980]: I1206 03:50:52.632302 4980 generic.go:334] "Generic (PLEG): container finished" podID="59a18727-c4b3-488f-ad90-08c32fd14878" containerID="554cf0ad092a2ea4bfc88306c34a4f23b7bd0388e5b01f77e23bf8470038e917" exitCode=0 Dec 06 03:50:52 crc kubenswrapper[4980]: I1206 03:50:52.632359 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-6pbbz" event={"ID":"59a18727-c4b3-488f-ad90-08c32fd14878","Type":"ContainerDied","Data":"554cf0ad092a2ea4bfc88306c34a4f23b7bd0388e5b01f77e23bf8470038e917"} Dec 06 03:50:52 crc kubenswrapper[4980]: I1206 03:50:52.632389 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-6pbbz" event={"ID":"59a18727-c4b3-488f-ad90-08c32fd14878","Type":"ContainerStarted","Data":"0e9314dededba865225a172203c5842f1b0d94a7faa663445e20ed522e5b2ed9"} Dec 06 03:50:53 crc kubenswrapper[4980]: I1206 03:50:53.640204 4980 generic.go:334] "Generic (PLEG): container finished" podID="59a18727-c4b3-488f-ad90-08c32fd14878" containerID="c492623b105b206deeb3af8ab6044e20ab7ff1f9618f49867d00290f843060e9" exitCode=0 Dec 06 03:50:53 crc kubenswrapper[4980]: I1206 03:50:53.640246 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-6pbbz" event={"ID":"59a18727-c4b3-488f-ad90-08c32fd14878","Type":"ContainerDied","Data":"c492623b105b206deeb3af8ab6044e20ab7ff1f9618f49867d00290f843060e9"} Dec 06 03:50:54 crc kubenswrapper[4980]: I1206 03:50:54.503563 4980 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/rabbitmq-server-0"] Dec 06 03:50:54 crc kubenswrapper[4980]: I1206 03:50:54.505083 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/rabbitmq-server-0" Dec 06 03:50:54 crc kubenswrapper[4980]: I1206 03:50:54.507847 4980 reflector.go:368] Caches populated for *v1.ConfigMap from object-"glance-kuttl-tests"/"rabbitmq-plugins-conf" Dec 06 03:50:54 crc kubenswrapper[4980]: I1206 03:50:54.508058 4980 reflector.go:368] Caches populated for *v1.ConfigMap from object-"glance-kuttl-tests"/"rabbitmq-server-conf" Dec 06 03:50:54 crc kubenswrapper[4980]: I1206 03:50:54.508231 4980 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"rabbitmq-erlang-cookie" Dec 06 03:50:54 crc kubenswrapper[4980]: I1206 03:50:54.508724 4980 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"rabbitmq-server-dockercfg-xk29w" Dec 06 03:50:54 crc kubenswrapper[4980]: I1206 03:50:54.509387 4980 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"rabbitmq-default-user" Dec 06 03:50:54 crc kubenswrapper[4980]: I1206 03:50:54.523154 4980 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/rabbitmq-server-0"] Dec 06 03:50:54 crc kubenswrapper[4980]: I1206 03:50:54.646960 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/92922f24-43cc-445d-a4ad-adcbf68b6636-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"92922f24-43cc-445d-a4ad-adcbf68b6636\") " pod="glance-kuttl-tests/rabbitmq-server-0" Dec 06 03:50:54 crc kubenswrapper[4980]: I1206 03:50:54.647033 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/92922f24-43cc-445d-a4ad-adcbf68b6636-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"92922f24-43cc-445d-a4ad-adcbf68b6636\") " pod="glance-kuttl-tests/rabbitmq-server-0" Dec 06 03:50:54 crc kubenswrapper[4980]: I1206 03:50:54.647311 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-b516b1da-4fb1-4ef8-94f5-c970f46c561f\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b516b1da-4fb1-4ef8-94f5-c970f46c561f\") pod \"rabbitmq-server-0\" (UID: \"92922f24-43cc-445d-a4ad-adcbf68b6636\") " pod="glance-kuttl-tests/rabbitmq-server-0" Dec 06 03:50:54 crc kubenswrapper[4980]: I1206 03:50:54.647343 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/92922f24-43cc-445d-a4ad-adcbf68b6636-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"92922f24-43cc-445d-a4ad-adcbf68b6636\") " pod="glance-kuttl-tests/rabbitmq-server-0" Dec 06 03:50:54 crc kubenswrapper[4980]: I1206 03:50:54.647377 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rmxx5\" (UniqueName: \"kubernetes.io/projected/92922f24-43cc-445d-a4ad-adcbf68b6636-kube-api-access-rmxx5\") pod \"rabbitmq-server-0\" (UID: \"92922f24-43cc-445d-a4ad-adcbf68b6636\") " pod="glance-kuttl-tests/rabbitmq-server-0" Dec 06 03:50:54 crc kubenswrapper[4980]: I1206 03:50:54.647775 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/92922f24-43cc-445d-a4ad-adcbf68b6636-pod-info\") pod \"rabbitmq-server-0\" (UID: \"92922f24-43cc-445d-a4ad-adcbf68b6636\") " pod="glance-kuttl-tests/rabbitmq-server-0" Dec 06 03:50:54 crc kubenswrapper[4980]: I1206 03:50:54.647860 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/92922f24-43cc-445d-a4ad-adcbf68b6636-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"92922f24-43cc-445d-a4ad-adcbf68b6636\") " pod="glance-kuttl-tests/rabbitmq-server-0" Dec 06 03:50:54 crc kubenswrapper[4980]: I1206 03:50:54.647887 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/92922f24-43cc-445d-a4ad-adcbf68b6636-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"92922f24-43cc-445d-a4ad-adcbf68b6636\") " pod="glance-kuttl-tests/rabbitmq-server-0" Dec 06 03:50:54 crc kubenswrapper[4980]: I1206 03:50:54.649068 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-6pbbz" event={"ID":"59a18727-c4b3-488f-ad90-08c32fd14878","Type":"ContainerStarted","Data":"a4d388f9cb1d4b77f17b182cd18b136373e0b06c4a84393ffd6f432e819be0f4"} Dec 06 03:50:54 crc kubenswrapper[4980]: I1206 03:50:54.675928 4980 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-6pbbz" podStartSLOduration=2.256154494 podStartE2EDuration="3.675901864s" podCreationTimestamp="2025-12-06 03:50:51 +0000 UTC" firstStartedPulling="2025-12-06 03:50:52.633674171 +0000 UTC m=+1011.878425442" lastFinishedPulling="2025-12-06 03:50:54.053421531 +0000 UTC m=+1013.298172812" observedRunningTime="2025-12-06 03:50:54.669813299 +0000 UTC m=+1013.914564580" watchObservedRunningTime="2025-12-06 03:50:54.675901864 +0000 UTC m=+1013.920653145" Dec 06 03:50:54 crc kubenswrapper[4980]: I1206 03:50:54.749009 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/92922f24-43cc-445d-a4ad-adcbf68b6636-pod-info\") pod \"rabbitmq-server-0\" (UID: \"92922f24-43cc-445d-a4ad-adcbf68b6636\") " pod="glance-kuttl-tests/rabbitmq-server-0" Dec 06 03:50:54 crc kubenswrapper[4980]: I1206 03:50:54.749071 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/92922f24-43cc-445d-a4ad-adcbf68b6636-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"92922f24-43cc-445d-a4ad-adcbf68b6636\") " pod="glance-kuttl-tests/rabbitmq-server-0" Dec 06 03:50:54 crc kubenswrapper[4980]: I1206 03:50:54.749094 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/92922f24-43cc-445d-a4ad-adcbf68b6636-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"92922f24-43cc-445d-a4ad-adcbf68b6636\") " pod="glance-kuttl-tests/rabbitmq-server-0" Dec 06 03:50:54 crc kubenswrapper[4980]: I1206 03:50:54.749123 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/92922f24-43cc-445d-a4ad-adcbf68b6636-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"92922f24-43cc-445d-a4ad-adcbf68b6636\") " pod="glance-kuttl-tests/rabbitmq-server-0" Dec 06 03:50:54 crc kubenswrapper[4980]: I1206 03:50:54.749175 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/92922f24-43cc-445d-a4ad-adcbf68b6636-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"92922f24-43cc-445d-a4ad-adcbf68b6636\") " pod="glance-kuttl-tests/rabbitmq-server-0" Dec 06 03:50:54 crc kubenswrapper[4980]: I1206 03:50:54.749222 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-b516b1da-4fb1-4ef8-94f5-c970f46c561f\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b516b1da-4fb1-4ef8-94f5-c970f46c561f\") pod \"rabbitmq-server-0\" (UID: \"92922f24-43cc-445d-a4ad-adcbf68b6636\") " pod="glance-kuttl-tests/rabbitmq-server-0" Dec 06 03:50:54 crc kubenswrapper[4980]: I1206 03:50:54.749248 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/92922f24-43cc-445d-a4ad-adcbf68b6636-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"92922f24-43cc-445d-a4ad-adcbf68b6636\") " pod="glance-kuttl-tests/rabbitmq-server-0" Dec 06 03:50:54 crc kubenswrapper[4980]: I1206 03:50:54.749278 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rmxx5\" (UniqueName: \"kubernetes.io/projected/92922f24-43cc-445d-a4ad-adcbf68b6636-kube-api-access-rmxx5\") pod \"rabbitmq-server-0\" (UID: \"92922f24-43cc-445d-a4ad-adcbf68b6636\") " pod="glance-kuttl-tests/rabbitmq-server-0" Dec 06 03:50:54 crc kubenswrapper[4980]: I1206 03:50:54.749713 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/92922f24-43cc-445d-a4ad-adcbf68b6636-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"92922f24-43cc-445d-a4ad-adcbf68b6636\") " pod="glance-kuttl-tests/rabbitmq-server-0" Dec 06 03:50:54 crc kubenswrapper[4980]: I1206 03:50:54.750012 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/92922f24-43cc-445d-a4ad-adcbf68b6636-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"92922f24-43cc-445d-a4ad-adcbf68b6636\") " pod="glance-kuttl-tests/rabbitmq-server-0" Dec 06 03:50:54 crc kubenswrapper[4980]: I1206 03:50:54.750436 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/92922f24-43cc-445d-a4ad-adcbf68b6636-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"92922f24-43cc-445d-a4ad-adcbf68b6636\") " pod="glance-kuttl-tests/rabbitmq-server-0" Dec 06 03:50:54 crc kubenswrapper[4980]: I1206 03:50:54.754889 4980 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 06 03:50:54 crc kubenswrapper[4980]: I1206 03:50:54.754931 4980 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-b516b1da-4fb1-4ef8-94f5-c970f46c561f\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b516b1da-4fb1-4ef8-94f5-c970f46c561f\") pod \"rabbitmq-server-0\" (UID: \"92922f24-43cc-445d-a4ad-adcbf68b6636\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/0b94ebdc439e281c0b3eba528bcdeb2de277f1ff6bec3e8daaee7e603e632490/globalmount\"" pod="glance-kuttl-tests/rabbitmq-server-0" Dec 06 03:50:54 crc kubenswrapper[4980]: I1206 03:50:54.755115 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/92922f24-43cc-445d-a4ad-adcbf68b6636-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"92922f24-43cc-445d-a4ad-adcbf68b6636\") " pod="glance-kuttl-tests/rabbitmq-server-0" Dec 06 03:50:54 crc kubenswrapper[4980]: I1206 03:50:54.755123 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/92922f24-43cc-445d-a4ad-adcbf68b6636-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"92922f24-43cc-445d-a4ad-adcbf68b6636\") " pod="glance-kuttl-tests/rabbitmq-server-0" Dec 06 03:50:54 crc kubenswrapper[4980]: I1206 03:50:54.762195 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/92922f24-43cc-445d-a4ad-adcbf68b6636-pod-info\") pod \"rabbitmq-server-0\" (UID: \"92922f24-43cc-445d-a4ad-adcbf68b6636\") " pod="glance-kuttl-tests/rabbitmq-server-0" Dec 06 03:50:54 crc kubenswrapper[4980]: I1206 03:50:54.777527 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-b516b1da-4fb1-4ef8-94f5-c970f46c561f\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b516b1da-4fb1-4ef8-94f5-c970f46c561f\") pod \"rabbitmq-server-0\" (UID: \"92922f24-43cc-445d-a4ad-adcbf68b6636\") " pod="glance-kuttl-tests/rabbitmq-server-0" Dec 06 03:50:54 crc kubenswrapper[4980]: I1206 03:50:54.778880 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rmxx5\" (UniqueName: \"kubernetes.io/projected/92922f24-43cc-445d-a4ad-adcbf68b6636-kube-api-access-rmxx5\") pod \"rabbitmq-server-0\" (UID: \"92922f24-43cc-445d-a4ad-adcbf68b6636\") " pod="glance-kuttl-tests/rabbitmq-server-0" Dec 06 03:50:54 crc kubenswrapper[4980]: I1206 03:50:54.825760 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/rabbitmq-server-0" Dec 06 03:50:55 crc kubenswrapper[4980]: I1206 03:50:55.317460 4980 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/rabbitmq-server-0"] Dec 06 03:50:55 crc kubenswrapper[4980]: I1206 03:50:55.655908 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/rabbitmq-server-0" event={"ID":"92922f24-43cc-445d-a4ad-adcbf68b6636","Type":"ContainerStarted","Data":"b0087109c2de9a11b65aa3cdd1263c4f78afa225f451574bc8125b3b0e0ca0a5"} Dec 06 03:50:57 crc kubenswrapper[4980]: I1206 03:50:57.231216 4980 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/keystone-operator-index-6gf95"] Dec 06 03:50:57 crc kubenswrapper[4980]: I1206 03:50:57.233976 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-index-6gf95" Dec 06 03:50:57 crc kubenswrapper[4980]: I1206 03:50:57.235882 4980 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"keystone-operator-index-dockercfg-69jtx" Dec 06 03:50:57 crc kubenswrapper[4980]: I1206 03:50:57.238249 4980 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-index-6gf95"] Dec 06 03:50:57 crc kubenswrapper[4980]: I1206 03:50:57.391213 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-56mdt\" (UniqueName: \"kubernetes.io/projected/322d87be-f8f1-4827-a296-dfbe881158d1-kube-api-access-56mdt\") pod \"keystone-operator-index-6gf95\" (UID: \"322d87be-f8f1-4827-a296-dfbe881158d1\") " pod="openstack-operators/keystone-operator-index-6gf95" Dec 06 03:50:57 crc kubenswrapper[4980]: I1206 03:50:57.492385 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-56mdt\" (UniqueName: \"kubernetes.io/projected/322d87be-f8f1-4827-a296-dfbe881158d1-kube-api-access-56mdt\") pod \"keystone-operator-index-6gf95\" (UID: \"322d87be-f8f1-4827-a296-dfbe881158d1\") " pod="openstack-operators/keystone-operator-index-6gf95" Dec 06 03:50:57 crc kubenswrapper[4980]: I1206 03:50:57.515707 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-56mdt\" (UniqueName: \"kubernetes.io/projected/322d87be-f8f1-4827-a296-dfbe881158d1-kube-api-access-56mdt\") pod \"keystone-operator-index-6gf95\" (UID: \"322d87be-f8f1-4827-a296-dfbe881158d1\") " pod="openstack-operators/keystone-operator-index-6gf95" Dec 06 03:50:57 crc kubenswrapper[4980]: I1206 03:50:57.554579 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-index-6gf95" Dec 06 03:50:57 crc kubenswrapper[4980]: I1206 03:50:57.973595 4980 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-index-6gf95"] Dec 06 03:50:57 crc kubenswrapper[4980]: W1206 03:50:57.982958 4980 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod322d87be_f8f1_4827_a296_dfbe881158d1.slice/crio-fc82cb2c9eb702f578726dfd877ef2815fe7848634d49d6c523f5d03c78c9b29 WatchSource:0}: Error finding container fc82cb2c9eb702f578726dfd877ef2815fe7848634d49d6c523f5d03c78c9b29: Status 404 returned error can't find the container with id fc82cb2c9eb702f578726dfd877ef2815fe7848634d49d6c523f5d03c78c9b29 Dec 06 03:50:58 crc kubenswrapper[4980]: I1206 03:50:58.683923 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-index-6gf95" event={"ID":"322d87be-f8f1-4827-a296-dfbe881158d1","Type":"ContainerStarted","Data":"fc82cb2c9eb702f578726dfd877ef2815fe7848634d49d6c523f5d03c78c9b29"} Dec 06 03:51:01 crc kubenswrapper[4980]: I1206 03:51:01.554459 4980 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-6pbbz" Dec 06 03:51:01 crc kubenswrapper[4980]: I1206 03:51:01.554819 4980 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-6pbbz" Dec 06 03:51:01 crc kubenswrapper[4980]: I1206 03:51:01.596778 4980 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-6pbbz" Dec 06 03:51:01 crc kubenswrapper[4980]: I1206 03:51:01.753845 4980 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-6pbbz" Dec 06 03:51:02 crc kubenswrapper[4980]: I1206 03:51:02.223733 4980 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/keystone-operator-index-6gf95"] Dec 06 03:51:03 crc kubenswrapper[4980]: I1206 03:51:03.033086 4980 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/keystone-operator-index-4jvwv"] Dec 06 03:51:03 crc kubenswrapper[4980]: I1206 03:51:03.035029 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-index-4jvwv" Dec 06 03:51:03 crc kubenswrapper[4980]: I1206 03:51:03.062030 4980 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-index-4jvwv"] Dec 06 03:51:03 crc kubenswrapper[4980]: I1206 03:51:03.196816 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-25d5b\" (UniqueName: \"kubernetes.io/projected/209b6261-c681-433d-884b-6edad0dc505d-kube-api-access-25d5b\") pod \"keystone-operator-index-4jvwv\" (UID: \"209b6261-c681-433d-884b-6edad0dc505d\") " pod="openstack-operators/keystone-operator-index-4jvwv" Dec 06 03:51:03 crc kubenswrapper[4980]: I1206 03:51:03.297888 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-25d5b\" (UniqueName: \"kubernetes.io/projected/209b6261-c681-433d-884b-6edad0dc505d-kube-api-access-25d5b\") pod \"keystone-operator-index-4jvwv\" (UID: \"209b6261-c681-433d-884b-6edad0dc505d\") " pod="openstack-operators/keystone-operator-index-4jvwv" Dec 06 03:51:03 crc kubenswrapper[4980]: I1206 03:51:03.324726 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-25d5b\" (UniqueName: \"kubernetes.io/projected/209b6261-c681-433d-884b-6edad0dc505d-kube-api-access-25d5b\") pod \"keystone-operator-index-4jvwv\" (UID: \"209b6261-c681-433d-884b-6edad0dc505d\") " pod="openstack-operators/keystone-operator-index-4jvwv" Dec 06 03:51:03 crc kubenswrapper[4980]: I1206 03:51:03.363678 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-index-4jvwv" Dec 06 03:51:04 crc kubenswrapper[4980]: I1206 03:51:04.227437 4980 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-index-4jvwv"] Dec 06 03:51:04 crc kubenswrapper[4980]: W1206 03:51:04.232429 4980 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod209b6261_c681_433d_884b_6edad0dc505d.slice/crio-40d4c545c8984c8a6c71ca595a703e30b5f896010ce47060e8a5e6cb67879ad1 WatchSource:0}: Error finding container 40d4c545c8984c8a6c71ca595a703e30b5f896010ce47060e8a5e6cb67879ad1: Status 404 returned error can't find the container with id 40d4c545c8984c8a6c71ca595a703e30b5f896010ce47060e8a5e6cb67879ad1 Dec 06 03:51:04 crc kubenswrapper[4980]: I1206 03:51:04.735267 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-index-4jvwv" event={"ID":"209b6261-c681-433d-884b-6edad0dc505d","Type":"ContainerStarted","Data":"40d4c545c8984c8a6c71ca595a703e30b5f896010ce47060e8a5e6cb67879ad1"} Dec 06 03:51:04 crc kubenswrapper[4980]: I1206 03:51:04.737707 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-index-6gf95" event={"ID":"322d87be-f8f1-4827-a296-dfbe881158d1","Type":"ContainerStarted","Data":"e4fcd2fd3edf81777c3cab7b649ceb9f805d7993732e2d087e5977018d85ffa6"} Dec 06 03:51:04 crc kubenswrapper[4980]: I1206 03:51:04.737923 4980 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/keystone-operator-index-6gf95" podUID="322d87be-f8f1-4827-a296-dfbe881158d1" containerName="registry-server" containerID="cri-o://e4fcd2fd3edf81777c3cab7b649ceb9f805d7993732e2d087e5977018d85ffa6" gracePeriod=2 Dec 06 03:51:04 crc kubenswrapper[4980]: I1206 03:51:04.739998 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/rabbitmq-server-0" event={"ID":"92922f24-43cc-445d-a4ad-adcbf68b6636","Type":"ContainerStarted","Data":"8d14297fc9e26de59e87c5641efc156518dd1526d321fda49138344216dca6fc"} Dec 06 03:51:04 crc kubenswrapper[4980]: I1206 03:51:04.761936 4980 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/keystone-operator-index-6gf95" podStartSLOduration=1.64041535 podStartE2EDuration="7.761910985s" podCreationTimestamp="2025-12-06 03:50:57 +0000 UTC" firstStartedPulling="2025-12-06 03:50:57.98611752 +0000 UTC m=+1017.230868791" lastFinishedPulling="2025-12-06 03:51:04.107613155 +0000 UTC m=+1023.352364426" observedRunningTime="2025-12-06 03:51:04.75720499 +0000 UTC m=+1024.001956281" watchObservedRunningTime="2025-12-06 03:51:04.761910985 +0000 UTC m=+1024.006662296" Dec 06 03:51:05 crc kubenswrapper[4980]: I1206 03:51:05.134946 4980 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-index-6gf95" Dec 06 03:51:05 crc kubenswrapper[4980]: I1206 03:51:05.223930 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-56mdt\" (UniqueName: \"kubernetes.io/projected/322d87be-f8f1-4827-a296-dfbe881158d1-kube-api-access-56mdt\") pod \"322d87be-f8f1-4827-a296-dfbe881158d1\" (UID: \"322d87be-f8f1-4827-a296-dfbe881158d1\") " Dec 06 03:51:05 crc kubenswrapper[4980]: I1206 03:51:05.230215 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/322d87be-f8f1-4827-a296-dfbe881158d1-kube-api-access-56mdt" (OuterVolumeSpecName: "kube-api-access-56mdt") pod "322d87be-f8f1-4827-a296-dfbe881158d1" (UID: "322d87be-f8f1-4827-a296-dfbe881158d1"). InnerVolumeSpecName "kube-api-access-56mdt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 03:51:05 crc kubenswrapper[4980]: I1206 03:51:05.325818 4980 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-56mdt\" (UniqueName: \"kubernetes.io/projected/322d87be-f8f1-4827-a296-dfbe881158d1-kube-api-access-56mdt\") on node \"crc\" DevicePath \"\"" Dec 06 03:51:05 crc kubenswrapper[4980]: I1206 03:51:05.748595 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-index-4jvwv" event={"ID":"209b6261-c681-433d-884b-6edad0dc505d","Type":"ContainerStarted","Data":"f9b362b2371b48e6692df9eca6b884d99a54a2c059431a1f11ba4c8bcccfcf02"} Dec 06 03:51:05 crc kubenswrapper[4980]: I1206 03:51:05.749789 4980 generic.go:334] "Generic (PLEG): container finished" podID="322d87be-f8f1-4827-a296-dfbe881158d1" containerID="e4fcd2fd3edf81777c3cab7b649ceb9f805d7993732e2d087e5977018d85ffa6" exitCode=0 Dec 06 03:51:05 crc kubenswrapper[4980]: I1206 03:51:05.749896 4980 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-index-6gf95" Dec 06 03:51:05 crc kubenswrapper[4980]: I1206 03:51:05.749906 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-index-6gf95" event={"ID":"322d87be-f8f1-4827-a296-dfbe881158d1","Type":"ContainerDied","Data":"e4fcd2fd3edf81777c3cab7b649ceb9f805d7993732e2d087e5977018d85ffa6"} Dec 06 03:51:05 crc kubenswrapper[4980]: I1206 03:51:05.749954 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-index-6gf95" event={"ID":"322d87be-f8f1-4827-a296-dfbe881158d1","Type":"ContainerDied","Data":"fc82cb2c9eb702f578726dfd877ef2815fe7848634d49d6c523f5d03c78c9b29"} Dec 06 03:51:05 crc kubenswrapper[4980]: I1206 03:51:05.749987 4980 scope.go:117] "RemoveContainer" containerID="e4fcd2fd3edf81777c3cab7b649ceb9f805d7993732e2d087e5977018d85ffa6" Dec 06 03:51:05 crc kubenswrapper[4980]: I1206 03:51:05.771559 4980 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/keystone-operator-index-4jvwv" podStartSLOduration=2.135945915 podStartE2EDuration="2.771536756s" podCreationTimestamp="2025-12-06 03:51:03 +0000 UTC" firstStartedPulling="2025-12-06 03:51:04.237841895 +0000 UTC m=+1023.482593166" lastFinishedPulling="2025-12-06 03:51:04.873432726 +0000 UTC m=+1024.118184007" observedRunningTime="2025-12-06 03:51:05.766564853 +0000 UTC m=+1025.011316144" watchObservedRunningTime="2025-12-06 03:51:05.771536756 +0000 UTC m=+1025.016288027" Dec 06 03:51:05 crc kubenswrapper[4980]: I1206 03:51:05.779733 4980 scope.go:117] "RemoveContainer" containerID="e4fcd2fd3edf81777c3cab7b649ceb9f805d7993732e2d087e5977018d85ffa6" Dec 06 03:51:05 crc kubenswrapper[4980]: E1206 03:51:05.780771 4980 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e4fcd2fd3edf81777c3cab7b649ceb9f805d7993732e2d087e5977018d85ffa6\": container with ID starting with e4fcd2fd3edf81777c3cab7b649ceb9f805d7993732e2d087e5977018d85ffa6 not found: ID does not exist" containerID="e4fcd2fd3edf81777c3cab7b649ceb9f805d7993732e2d087e5977018d85ffa6" Dec 06 03:51:05 crc kubenswrapper[4980]: I1206 03:51:05.780837 4980 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e4fcd2fd3edf81777c3cab7b649ceb9f805d7993732e2d087e5977018d85ffa6"} err="failed to get container status \"e4fcd2fd3edf81777c3cab7b649ceb9f805d7993732e2d087e5977018d85ffa6\": rpc error: code = NotFound desc = could not find container \"e4fcd2fd3edf81777c3cab7b649ceb9f805d7993732e2d087e5977018d85ffa6\": container with ID starting with e4fcd2fd3edf81777c3cab7b649ceb9f805d7993732e2d087e5977018d85ffa6 not found: ID does not exist" Dec 06 03:51:05 crc kubenswrapper[4980]: I1206 03:51:05.799149 4980 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/keystone-operator-index-6gf95"] Dec 06 03:51:05 crc kubenswrapper[4980]: I1206 03:51:05.805850 4980 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/keystone-operator-index-6gf95"] Dec 06 03:51:07 crc kubenswrapper[4980]: I1206 03:51:07.194149 4980 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="322d87be-f8f1-4827-a296-dfbe881158d1" path="/var/lib/kubelet/pods/322d87be-f8f1-4827-a296-dfbe881158d1/volumes" Dec 06 03:51:09 crc kubenswrapper[4980]: I1206 03:51:09.428604 4980 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-6pbbz"] Dec 06 03:51:09 crc kubenswrapper[4980]: I1206 03:51:09.429285 4980 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-6pbbz" podUID="59a18727-c4b3-488f-ad90-08c32fd14878" containerName="registry-server" containerID="cri-o://a4d388f9cb1d4b77f17b182cd18b136373e0b06c4a84393ffd6f432e819be0f4" gracePeriod=2 Dec 06 03:51:11 crc kubenswrapper[4980]: E1206 03:51:11.554843 4980 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of a4d388f9cb1d4b77f17b182cd18b136373e0b06c4a84393ffd6f432e819be0f4 is running failed: container process not found" containerID="a4d388f9cb1d4b77f17b182cd18b136373e0b06c4a84393ffd6f432e819be0f4" cmd=["grpc_health_probe","-addr=:50051"] Dec 06 03:51:11 crc kubenswrapper[4980]: E1206 03:51:11.555850 4980 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of a4d388f9cb1d4b77f17b182cd18b136373e0b06c4a84393ffd6f432e819be0f4 is running failed: container process not found" containerID="a4d388f9cb1d4b77f17b182cd18b136373e0b06c4a84393ffd6f432e819be0f4" cmd=["grpc_health_probe","-addr=:50051"] Dec 06 03:51:11 crc kubenswrapper[4980]: E1206 03:51:11.556292 4980 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of a4d388f9cb1d4b77f17b182cd18b136373e0b06c4a84393ffd6f432e819be0f4 is running failed: container process not found" containerID="a4d388f9cb1d4b77f17b182cd18b136373e0b06c4a84393ffd6f432e819be0f4" cmd=["grpc_health_probe","-addr=:50051"] Dec 06 03:51:11 crc kubenswrapper[4980]: E1206 03:51:11.556328 4980 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of a4d388f9cb1d4b77f17b182cd18b136373e0b06c4a84393ffd6f432e819be0f4 is running failed: container process not found" probeType="Readiness" pod="openshift-marketplace/redhat-marketplace-6pbbz" podUID="59a18727-c4b3-488f-ad90-08c32fd14878" containerName="registry-server" Dec 06 03:51:11 crc kubenswrapper[4980]: I1206 03:51:11.801347 4980 generic.go:334] "Generic (PLEG): container finished" podID="59a18727-c4b3-488f-ad90-08c32fd14878" containerID="a4d388f9cb1d4b77f17b182cd18b136373e0b06c4a84393ffd6f432e819be0f4" exitCode=0 Dec 06 03:51:11 crc kubenswrapper[4980]: I1206 03:51:11.801408 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-6pbbz" event={"ID":"59a18727-c4b3-488f-ad90-08c32fd14878","Type":"ContainerDied","Data":"a4d388f9cb1d4b77f17b182cd18b136373e0b06c4a84393ffd6f432e819be0f4"} Dec 06 03:51:12 crc kubenswrapper[4980]: I1206 03:51:12.558013 4980 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-6pbbz" Dec 06 03:51:12 crc kubenswrapper[4980]: I1206 03:51:12.649161 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/59a18727-c4b3-488f-ad90-08c32fd14878-catalog-content\") pod \"59a18727-c4b3-488f-ad90-08c32fd14878\" (UID: \"59a18727-c4b3-488f-ad90-08c32fd14878\") " Dec 06 03:51:12 crc kubenswrapper[4980]: I1206 03:51:12.649294 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w588x\" (UniqueName: \"kubernetes.io/projected/59a18727-c4b3-488f-ad90-08c32fd14878-kube-api-access-w588x\") pod \"59a18727-c4b3-488f-ad90-08c32fd14878\" (UID: \"59a18727-c4b3-488f-ad90-08c32fd14878\") " Dec 06 03:51:12 crc kubenswrapper[4980]: I1206 03:51:12.649378 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/59a18727-c4b3-488f-ad90-08c32fd14878-utilities\") pod \"59a18727-c4b3-488f-ad90-08c32fd14878\" (UID: \"59a18727-c4b3-488f-ad90-08c32fd14878\") " Dec 06 03:51:12 crc kubenswrapper[4980]: I1206 03:51:12.650446 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/59a18727-c4b3-488f-ad90-08c32fd14878-utilities" (OuterVolumeSpecName: "utilities") pod "59a18727-c4b3-488f-ad90-08c32fd14878" (UID: "59a18727-c4b3-488f-ad90-08c32fd14878"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 03:51:12 crc kubenswrapper[4980]: I1206 03:51:12.655696 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/59a18727-c4b3-488f-ad90-08c32fd14878-kube-api-access-w588x" (OuterVolumeSpecName: "kube-api-access-w588x") pod "59a18727-c4b3-488f-ad90-08c32fd14878" (UID: "59a18727-c4b3-488f-ad90-08c32fd14878"). InnerVolumeSpecName "kube-api-access-w588x". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 03:51:12 crc kubenswrapper[4980]: I1206 03:51:12.678481 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/59a18727-c4b3-488f-ad90-08c32fd14878-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "59a18727-c4b3-488f-ad90-08c32fd14878" (UID: "59a18727-c4b3-488f-ad90-08c32fd14878"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 03:51:12 crc kubenswrapper[4980]: I1206 03:51:12.752009 4980 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/59a18727-c4b3-488f-ad90-08c32fd14878-utilities\") on node \"crc\" DevicePath \"\"" Dec 06 03:51:12 crc kubenswrapper[4980]: I1206 03:51:12.752073 4980 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/59a18727-c4b3-488f-ad90-08c32fd14878-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 06 03:51:12 crc kubenswrapper[4980]: I1206 03:51:12.752093 4980 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w588x\" (UniqueName: \"kubernetes.io/projected/59a18727-c4b3-488f-ad90-08c32fd14878-kube-api-access-w588x\") on node \"crc\" DevicePath \"\"" Dec 06 03:51:12 crc kubenswrapper[4980]: I1206 03:51:12.811859 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-6pbbz" event={"ID":"59a18727-c4b3-488f-ad90-08c32fd14878","Type":"ContainerDied","Data":"0e9314dededba865225a172203c5842f1b0d94a7faa663445e20ed522e5b2ed9"} Dec 06 03:51:12 crc kubenswrapper[4980]: I1206 03:51:12.811961 4980 scope.go:117] "RemoveContainer" containerID="a4d388f9cb1d4b77f17b182cd18b136373e0b06c4a84393ffd6f432e819be0f4" Dec 06 03:51:12 crc kubenswrapper[4980]: I1206 03:51:12.811983 4980 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-6pbbz" Dec 06 03:51:12 crc kubenswrapper[4980]: I1206 03:51:12.830977 4980 scope.go:117] "RemoveContainer" containerID="c492623b105b206deeb3af8ab6044e20ab7ff1f9618f49867d00290f843060e9" Dec 06 03:51:12 crc kubenswrapper[4980]: I1206 03:51:12.853452 4980 scope.go:117] "RemoveContainer" containerID="554cf0ad092a2ea4bfc88306c34a4f23b7bd0388e5b01f77e23bf8470038e917" Dec 06 03:51:12 crc kubenswrapper[4980]: I1206 03:51:12.856093 4980 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-6pbbz"] Dec 06 03:51:12 crc kubenswrapper[4980]: I1206 03:51:12.861002 4980 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-6pbbz"] Dec 06 03:51:13 crc kubenswrapper[4980]: I1206 03:51:13.194641 4980 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="59a18727-c4b3-488f-ad90-08c32fd14878" path="/var/lib/kubelet/pods/59a18727-c4b3-488f-ad90-08c32fd14878/volumes" Dec 06 03:51:13 crc kubenswrapper[4980]: I1206 03:51:13.364769 4980 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/keystone-operator-index-4jvwv" Dec 06 03:51:13 crc kubenswrapper[4980]: I1206 03:51:13.364823 4980 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/keystone-operator-index-4jvwv" Dec 06 03:51:13 crc kubenswrapper[4980]: I1206 03:51:13.404665 4980 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/keystone-operator-index-4jvwv" Dec 06 03:51:13 crc kubenswrapper[4980]: I1206 03:51:13.860605 4980 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/keystone-operator-index-4jvwv" Dec 06 03:51:14 crc kubenswrapper[4980]: I1206 03:51:14.241315 4980 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-zvj4s"] Dec 06 03:51:14 crc kubenswrapper[4980]: E1206 03:51:14.241963 4980 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="59a18727-c4b3-488f-ad90-08c32fd14878" containerName="extract-utilities" Dec 06 03:51:14 crc kubenswrapper[4980]: I1206 03:51:14.241981 4980 state_mem.go:107] "Deleted CPUSet assignment" podUID="59a18727-c4b3-488f-ad90-08c32fd14878" containerName="extract-utilities" Dec 06 03:51:14 crc kubenswrapper[4980]: E1206 03:51:14.242012 4980 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="59a18727-c4b3-488f-ad90-08c32fd14878" containerName="registry-server" Dec 06 03:51:14 crc kubenswrapper[4980]: I1206 03:51:14.242022 4980 state_mem.go:107] "Deleted CPUSet assignment" podUID="59a18727-c4b3-488f-ad90-08c32fd14878" containerName="registry-server" Dec 06 03:51:14 crc kubenswrapper[4980]: E1206 03:51:14.242036 4980 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="322d87be-f8f1-4827-a296-dfbe881158d1" containerName="registry-server" Dec 06 03:51:14 crc kubenswrapper[4980]: I1206 03:51:14.242044 4980 state_mem.go:107] "Deleted CPUSet assignment" podUID="322d87be-f8f1-4827-a296-dfbe881158d1" containerName="registry-server" Dec 06 03:51:14 crc kubenswrapper[4980]: E1206 03:51:14.242057 4980 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="59a18727-c4b3-488f-ad90-08c32fd14878" containerName="extract-content" Dec 06 03:51:14 crc kubenswrapper[4980]: I1206 03:51:14.242065 4980 state_mem.go:107] "Deleted CPUSet assignment" podUID="59a18727-c4b3-488f-ad90-08c32fd14878" containerName="extract-content" Dec 06 03:51:14 crc kubenswrapper[4980]: I1206 03:51:14.242210 4980 memory_manager.go:354] "RemoveStaleState removing state" podUID="59a18727-c4b3-488f-ad90-08c32fd14878" containerName="registry-server" Dec 06 03:51:14 crc kubenswrapper[4980]: I1206 03:51:14.242228 4980 memory_manager.go:354] "RemoveStaleState removing state" podUID="322d87be-f8f1-4827-a296-dfbe881158d1" containerName="registry-server" Dec 06 03:51:14 crc kubenswrapper[4980]: I1206 03:51:14.243865 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-zvj4s" Dec 06 03:51:14 crc kubenswrapper[4980]: I1206 03:51:14.257980 4980 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-zvj4s"] Dec 06 03:51:14 crc kubenswrapper[4980]: I1206 03:51:14.377838 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/483776ab-7b20-420b-94f6-fdd02f3fc912-utilities\") pod \"certified-operators-zvj4s\" (UID: \"483776ab-7b20-420b-94f6-fdd02f3fc912\") " pod="openshift-marketplace/certified-operators-zvj4s" Dec 06 03:51:14 crc kubenswrapper[4980]: I1206 03:51:14.377907 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/483776ab-7b20-420b-94f6-fdd02f3fc912-catalog-content\") pod \"certified-operators-zvj4s\" (UID: \"483776ab-7b20-420b-94f6-fdd02f3fc912\") " pod="openshift-marketplace/certified-operators-zvj4s" Dec 06 03:51:14 crc kubenswrapper[4980]: I1206 03:51:14.377971 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z95k7\" (UniqueName: \"kubernetes.io/projected/483776ab-7b20-420b-94f6-fdd02f3fc912-kube-api-access-z95k7\") pod \"certified-operators-zvj4s\" (UID: \"483776ab-7b20-420b-94f6-fdd02f3fc912\") " pod="openshift-marketplace/certified-operators-zvj4s" Dec 06 03:51:14 crc kubenswrapper[4980]: I1206 03:51:14.479951 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/483776ab-7b20-420b-94f6-fdd02f3fc912-catalog-content\") pod \"certified-operators-zvj4s\" (UID: \"483776ab-7b20-420b-94f6-fdd02f3fc912\") " pod="openshift-marketplace/certified-operators-zvj4s" Dec 06 03:51:14 crc kubenswrapper[4980]: I1206 03:51:14.480008 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/483776ab-7b20-420b-94f6-fdd02f3fc912-catalog-content\") pod \"certified-operators-zvj4s\" (UID: \"483776ab-7b20-420b-94f6-fdd02f3fc912\") " pod="openshift-marketplace/certified-operators-zvj4s" Dec 06 03:51:14 crc kubenswrapper[4980]: I1206 03:51:14.480077 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z95k7\" (UniqueName: \"kubernetes.io/projected/483776ab-7b20-420b-94f6-fdd02f3fc912-kube-api-access-z95k7\") pod \"certified-operators-zvj4s\" (UID: \"483776ab-7b20-420b-94f6-fdd02f3fc912\") " pod="openshift-marketplace/certified-operators-zvj4s" Dec 06 03:51:14 crc kubenswrapper[4980]: I1206 03:51:14.480443 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/483776ab-7b20-420b-94f6-fdd02f3fc912-utilities\") pod \"certified-operators-zvj4s\" (UID: \"483776ab-7b20-420b-94f6-fdd02f3fc912\") " pod="openshift-marketplace/certified-operators-zvj4s" Dec 06 03:51:14 crc kubenswrapper[4980]: I1206 03:51:14.480796 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/483776ab-7b20-420b-94f6-fdd02f3fc912-utilities\") pod \"certified-operators-zvj4s\" (UID: \"483776ab-7b20-420b-94f6-fdd02f3fc912\") " pod="openshift-marketplace/certified-operators-zvj4s" Dec 06 03:51:14 crc kubenswrapper[4980]: I1206 03:51:14.507727 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z95k7\" (UniqueName: \"kubernetes.io/projected/483776ab-7b20-420b-94f6-fdd02f3fc912-kube-api-access-z95k7\") pod \"certified-operators-zvj4s\" (UID: \"483776ab-7b20-420b-94f6-fdd02f3fc912\") " pod="openshift-marketplace/certified-operators-zvj4s" Dec 06 03:51:14 crc kubenswrapper[4980]: I1206 03:51:14.561284 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-zvj4s" Dec 06 03:51:15 crc kubenswrapper[4980]: I1206 03:51:15.106648 4980 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-zvj4s"] Dec 06 03:51:15 crc kubenswrapper[4980]: W1206 03:51:15.116404 4980 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod483776ab_7b20_420b_94f6_fdd02f3fc912.slice/crio-9514dd72bb7c856ad8bf9b9a64ef90aed5ccc9ba59222973de547e8d4863a927 WatchSource:0}: Error finding container 9514dd72bb7c856ad8bf9b9a64ef90aed5ccc9ba59222973de547e8d4863a927: Status 404 returned error can't find the container with id 9514dd72bb7c856ad8bf9b9a64ef90aed5ccc9ba59222973de547e8d4863a927 Dec 06 03:51:15 crc kubenswrapper[4980]: I1206 03:51:15.845861 4980 generic.go:334] "Generic (PLEG): container finished" podID="483776ab-7b20-420b-94f6-fdd02f3fc912" containerID="cfe03737916282aae1effa54360d48ac1b0fdd879cda738f92fbe5de23c71b6e" exitCode=0 Dec 06 03:51:15 crc kubenswrapper[4980]: I1206 03:51:15.845957 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zvj4s" event={"ID":"483776ab-7b20-420b-94f6-fdd02f3fc912","Type":"ContainerDied","Data":"cfe03737916282aae1effa54360d48ac1b0fdd879cda738f92fbe5de23c71b6e"} Dec 06 03:51:15 crc kubenswrapper[4980]: I1206 03:51:15.846254 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zvj4s" event={"ID":"483776ab-7b20-420b-94f6-fdd02f3fc912","Type":"ContainerStarted","Data":"9514dd72bb7c856ad8bf9b9a64ef90aed5ccc9ba59222973de547e8d4863a927"} Dec 06 03:51:16 crc kubenswrapper[4980]: I1206 03:51:16.861221 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zvj4s" event={"ID":"483776ab-7b20-420b-94f6-fdd02f3fc912","Type":"ContainerStarted","Data":"fc2b66dd2ededd04478691ea1f398a4b1f2ec4149bb459799798582c8f8d1df8"} Dec 06 03:51:17 crc kubenswrapper[4980]: I1206 03:51:17.871079 4980 generic.go:334] "Generic (PLEG): container finished" podID="483776ab-7b20-420b-94f6-fdd02f3fc912" containerID="fc2b66dd2ededd04478691ea1f398a4b1f2ec4149bb459799798582c8f8d1df8" exitCode=0 Dec 06 03:51:17 crc kubenswrapper[4980]: I1206 03:51:17.871140 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zvj4s" event={"ID":"483776ab-7b20-420b-94f6-fdd02f3fc912","Type":"ContainerDied","Data":"fc2b66dd2ededd04478691ea1f398a4b1f2ec4149bb459799798582c8f8d1df8"} Dec 06 03:51:18 crc kubenswrapper[4980]: I1206 03:51:18.880632 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zvj4s" event={"ID":"483776ab-7b20-420b-94f6-fdd02f3fc912","Type":"ContainerStarted","Data":"dbfa1cd63a625381af83c9ed23757111a17b4c44671bdded826fab31160cf168"} Dec 06 03:51:24 crc kubenswrapper[4980]: I1206 03:51:24.562147 4980 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-zvj4s" Dec 06 03:51:24 crc kubenswrapper[4980]: I1206 03:51:24.562429 4980 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-zvj4s" Dec 06 03:51:24 crc kubenswrapper[4980]: I1206 03:51:24.604499 4980 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-zvj4s" Dec 06 03:51:24 crc kubenswrapper[4980]: I1206 03:51:24.622570 4980 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-zvj4s" podStartSLOduration=8.204441416 podStartE2EDuration="10.622530559s" podCreationTimestamp="2025-12-06 03:51:14 +0000 UTC" firstStartedPulling="2025-12-06 03:51:15.848768053 +0000 UTC m=+1035.093519364" lastFinishedPulling="2025-12-06 03:51:18.266857216 +0000 UTC m=+1037.511608507" observedRunningTime="2025-12-06 03:51:18.904896075 +0000 UTC m=+1038.149647346" watchObservedRunningTime="2025-12-06 03:51:24.622530559 +0000 UTC m=+1043.867281830" Dec 06 03:51:24 crc kubenswrapper[4980]: I1206 03:51:24.973256 4980 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-zvj4s" Dec 06 03:51:25 crc kubenswrapper[4980]: I1206 03:51:25.012392 4980 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-zvj4s"] Dec 06 03:51:26 crc kubenswrapper[4980]: I1206 03:51:26.944410 4980 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-zvj4s" podUID="483776ab-7b20-420b-94f6-fdd02f3fc912" containerName="registry-server" containerID="cri-o://dbfa1cd63a625381af83c9ed23757111a17b4c44671bdded826fab31160cf168" gracePeriod=2 Dec 06 03:51:28 crc kubenswrapper[4980]: I1206 03:51:28.468736 4980 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-zvj4s" Dec 06 03:51:28 crc kubenswrapper[4980]: I1206 03:51:28.626990 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/483776ab-7b20-420b-94f6-fdd02f3fc912-catalog-content\") pod \"483776ab-7b20-420b-94f6-fdd02f3fc912\" (UID: \"483776ab-7b20-420b-94f6-fdd02f3fc912\") " Dec 06 03:51:28 crc kubenswrapper[4980]: I1206 03:51:28.627039 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/483776ab-7b20-420b-94f6-fdd02f3fc912-utilities\") pod \"483776ab-7b20-420b-94f6-fdd02f3fc912\" (UID: \"483776ab-7b20-420b-94f6-fdd02f3fc912\") " Dec 06 03:51:28 crc kubenswrapper[4980]: I1206 03:51:28.627069 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z95k7\" (UniqueName: \"kubernetes.io/projected/483776ab-7b20-420b-94f6-fdd02f3fc912-kube-api-access-z95k7\") pod \"483776ab-7b20-420b-94f6-fdd02f3fc912\" (UID: \"483776ab-7b20-420b-94f6-fdd02f3fc912\") " Dec 06 03:51:28 crc kubenswrapper[4980]: I1206 03:51:28.628296 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/483776ab-7b20-420b-94f6-fdd02f3fc912-utilities" (OuterVolumeSpecName: "utilities") pod "483776ab-7b20-420b-94f6-fdd02f3fc912" (UID: "483776ab-7b20-420b-94f6-fdd02f3fc912"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 03:51:28 crc kubenswrapper[4980]: I1206 03:51:28.636902 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/483776ab-7b20-420b-94f6-fdd02f3fc912-kube-api-access-z95k7" (OuterVolumeSpecName: "kube-api-access-z95k7") pod "483776ab-7b20-420b-94f6-fdd02f3fc912" (UID: "483776ab-7b20-420b-94f6-fdd02f3fc912"). InnerVolumeSpecName "kube-api-access-z95k7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 03:51:28 crc kubenswrapper[4980]: I1206 03:51:28.675239 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/483776ab-7b20-420b-94f6-fdd02f3fc912-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "483776ab-7b20-420b-94f6-fdd02f3fc912" (UID: "483776ab-7b20-420b-94f6-fdd02f3fc912"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 03:51:28 crc kubenswrapper[4980]: I1206 03:51:28.729991 4980 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/483776ab-7b20-420b-94f6-fdd02f3fc912-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 06 03:51:28 crc kubenswrapper[4980]: I1206 03:51:28.730031 4980 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/483776ab-7b20-420b-94f6-fdd02f3fc912-utilities\") on node \"crc\" DevicePath \"\"" Dec 06 03:51:28 crc kubenswrapper[4980]: I1206 03:51:28.730044 4980 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z95k7\" (UniqueName: \"kubernetes.io/projected/483776ab-7b20-420b-94f6-fdd02f3fc912-kube-api-access-z95k7\") on node \"crc\" DevicePath \"\"" Dec 06 03:51:28 crc kubenswrapper[4980]: I1206 03:51:28.987296 4980 generic.go:334] "Generic (PLEG): container finished" podID="483776ab-7b20-420b-94f6-fdd02f3fc912" containerID="dbfa1cd63a625381af83c9ed23757111a17b4c44671bdded826fab31160cf168" exitCode=0 Dec 06 03:51:28 crc kubenswrapper[4980]: I1206 03:51:28.987399 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zvj4s" event={"ID":"483776ab-7b20-420b-94f6-fdd02f3fc912","Type":"ContainerDied","Data":"dbfa1cd63a625381af83c9ed23757111a17b4c44671bdded826fab31160cf168"} Dec 06 03:51:28 crc kubenswrapper[4980]: I1206 03:51:28.987446 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zvj4s" event={"ID":"483776ab-7b20-420b-94f6-fdd02f3fc912","Type":"ContainerDied","Data":"9514dd72bb7c856ad8bf9b9a64ef90aed5ccc9ba59222973de547e8d4863a927"} Dec 06 03:51:28 crc kubenswrapper[4980]: I1206 03:51:28.987481 4980 scope.go:117] "RemoveContainer" containerID="dbfa1cd63a625381af83c9ed23757111a17b4c44671bdded826fab31160cf168" Dec 06 03:51:28 crc kubenswrapper[4980]: I1206 03:51:28.987478 4980 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-zvj4s" Dec 06 03:51:29 crc kubenswrapper[4980]: I1206 03:51:29.018856 4980 scope.go:117] "RemoveContainer" containerID="fc2b66dd2ededd04478691ea1f398a4b1f2ec4149bb459799798582c8f8d1df8" Dec 06 03:51:29 crc kubenswrapper[4980]: I1206 03:51:29.046583 4980 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-zvj4s"] Dec 06 03:51:29 crc kubenswrapper[4980]: I1206 03:51:29.052161 4980 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-zvj4s"] Dec 06 03:51:29 crc kubenswrapper[4980]: I1206 03:51:29.053057 4980 scope.go:117] "RemoveContainer" containerID="cfe03737916282aae1effa54360d48ac1b0fdd879cda738f92fbe5de23c71b6e" Dec 06 03:51:29 crc kubenswrapper[4980]: I1206 03:51:29.085886 4980 scope.go:117] "RemoveContainer" containerID="dbfa1cd63a625381af83c9ed23757111a17b4c44671bdded826fab31160cf168" Dec 06 03:51:29 crc kubenswrapper[4980]: E1206 03:51:29.086548 4980 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dbfa1cd63a625381af83c9ed23757111a17b4c44671bdded826fab31160cf168\": container with ID starting with dbfa1cd63a625381af83c9ed23757111a17b4c44671bdded826fab31160cf168 not found: ID does not exist" containerID="dbfa1cd63a625381af83c9ed23757111a17b4c44671bdded826fab31160cf168" Dec 06 03:51:29 crc kubenswrapper[4980]: I1206 03:51:29.086598 4980 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dbfa1cd63a625381af83c9ed23757111a17b4c44671bdded826fab31160cf168"} err="failed to get container status \"dbfa1cd63a625381af83c9ed23757111a17b4c44671bdded826fab31160cf168\": rpc error: code = NotFound desc = could not find container \"dbfa1cd63a625381af83c9ed23757111a17b4c44671bdded826fab31160cf168\": container with ID starting with dbfa1cd63a625381af83c9ed23757111a17b4c44671bdded826fab31160cf168 not found: ID does not exist" Dec 06 03:51:29 crc kubenswrapper[4980]: I1206 03:51:29.086631 4980 scope.go:117] "RemoveContainer" containerID="fc2b66dd2ededd04478691ea1f398a4b1f2ec4149bb459799798582c8f8d1df8" Dec 06 03:51:29 crc kubenswrapper[4980]: E1206 03:51:29.086926 4980 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fc2b66dd2ededd04478691ea1f398a4b1f2ec4149bb459799798582c8f8d1df8\": container with ID starting with fc2b66dd2ededd04478691ea1f398a4b1f2ec4149bb459799798582c8f8d1df8 not found: ID does not exist" containerID="fc2b66dd2ededd04478691ea1f398a4b1f2ec4149bb459799798582c8f8d1df8" Dec 06 03:51:29 crc kubenswrapper[4980]: I1206 03:51:29.086953 4980 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fc2b66dd2ededd04478691ea1f398a4b1f2ec4149bb459799798582c8f8d1df8"} err="failed to get container status \"fc2b66dd2ededd04478691ea1f398a4b1f2ec4149bb459799798582c8f8d1df8\": rpc error: code = NotFound desc = could not find container \"fc2b66dd2ededd04478691ea1f398a4b1f2ec4149bb459799798582c8f8d1df8\": container with ID starting with fc2b66dd2ededd04478691ea1f398a4b1f2ec4149bb459799798582c8f8d1df8 not found: ID does not exist" Dec 06 03:51:29 crc kubenswrapper[4980]: I1206 03:51:29.086970 4980 scope.go:117] "RemoveContainer" containerID="cfe03737916282aae1effa54360d48ac1b0fdd879cda738f92fbe5de23c71b6e" Dec 06 03:51:29 crc kubenswrapper[4980]: E1206 03:51:29.087357 4980 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cfe03737916282aae1effa54360d48ac1b0fdd879cda738f92fbe5de23c71b6e\": container with ID starting with cfe03737916282aae1effa54360d48ac1b0fdd879cda738f92fbe5de23c71b6e not found: ID does not exist" containerID="cfe03737916282aae1effa54360d48ac1b0fdd879cda738f92fbe5de23c71b6e" Dec 06 03:51:29 crc kubenswrapper[4980]: I1206 03:51:29.087440 4980 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cfe03737916282aae1effa54360d48ac1b0fdd879cda738f92fbe5de23c71b6e"} err="failed to get container status \"cfe03737916282aae1effa54360d48ac1b0fdd879cda738f92fbe5de23c71b6e\": rpc error: code = NotFound desc = could not find container \"cfe03737916282aae1effa54360d48ac1b0fdd879cda738f92fbe5de23c71b6e\": container with ID starting with cfe03737916282aae1effa54360d48ac1b0fdd879cda738f92fbe5de23c71b6e not found: ID does not exist" Dec 06 03:51:29 crc kubenswrapper[4980]: I1206 03:51:29.192176 4980 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="483776ab-7b20-420b-94f6-fdd02f3fc912" path="/var/lib/kubelet/pods/483776ab-7b20-420b-94f6-fdd02f3fc912/volumes" Dec 06 03:51:30 crc kubenswrapper[4980]: I1206 03:51:30.094913 4980 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/673e18c374f427d47ebf1299aef3a4ea1aab2302d0e987986af30e26212ztjs"] Dec 06 03:51:30 crc kubenswrapper[4980]: E1206 03:51:30.095237 4980 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="483776ab-7b20-420b-94f6-fdd02f3fc912" containerName="extract-content" Dec 06 03:51:30 crc kubenswrapper[4980]: I1206 03:51:30.095252 4980 state_mem.go:107] "Deleted CPUSet assignment" podUID="483776ab-7b20-420b-94f6-fdd02f3fc912" containerName="extract-content" Dec 06 03:51:30 crc kubenswrapper[4980]: E1206 03:51:30.095272 4980 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="483776ab-7b20-420b-94f6-fdd02f3fc912" containerName="extract-utilities" Dec 06 03:51:30 crc kubenswrapper[4980]: I1206 03:51:30.095281 4980 state_mem.go:107] "Deleted CPUSet assignment" podUID="483776ab-7b20-420b-94f6-fdd02f3fc912" containerName="extract-utilities" Dec 06 03:51:30 crc kubenswrapper[4980]: E1206 03:51:30.095294 4980 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="483776ab-7b20-420b-94f6-fdd02f3fc912" containerName="registry-server" Dec 06 03:51:30 crc kubenswrapper[4980]: I1206 03:51:30.095303 4980 state_mem.go:107] "Deleted CPUSet assignment" podUID="483776ab-7b20-420b-94f6-fdd02f3fc912" containerName="registry-server" Dec 06 03:51:30 crc kubenswrapper[4980]: I1206 03:51:30.095455 4980 memory_manager.go:354] "RemoveStaleState removing state" podUID="483776ab-7b20-420b-94f6-fdd02f3fc912" containerName="registry-server" Dec 06 03:51:30 crc kubenswrapper[4980]: I1206 03:51:30.096487 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/673e18c374f427d47ebf1299aef3a4ea1aab2302d0e987986af30e26212ztjs" Dec 06 03:51:30 crc kubenswrapper[4980]: I1206 03:51:30.099916 4980 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"default-dockercfg-brs8h" Dec 06 03:51:30 crc kubenswrapper[4980]: I1206 03:51:30.105728 4980 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/673e18c374f427d47ebf1299aef3a4ea1aab2302d0e987986af30e26212ztjs"] Dec 06 03:51:30 crc kubenswrapper[4980]: I1206 03:51:30.161734 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5qpk4\" (UniqueName: \"kubernetes.io/projected/7171eff1-10af-4bab-b43b-02368608ee00-kube-api-access-5qpk4\") pod \"673e18c374f427d47ebf1299aef3a4ea1aab2302d0e987986af30e26212ztjs\" (UID: \"7171eff1-10af-4bab-b43b-02368608ee00\") " pod="openstack-operators/673e18c374f427d47ebf1299aef3a4ea1aab2302d0e987986af30e26212ztjs" Dec 06 03:51:30 crc kubenswrapper[4980]: I1206 03:51:30.161874 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/7171eff1-10af-4bab-b43b-02368608ee00-bundle\") pod \"673e18c374f427d47ebf1299aef3a4ea1aab2302d0e987986af30e26212ztjs\" (UID: \"7171eff1-10af-4bab-b43b-02368608ee00\") " pod="openstack-operators/673e18c374f427d47ebf1299aef3a4ea1aab2302d0e987986af30e26212ztjs" Dec 06 03:51:30 crc kubenswrapper[4980]: I1206 03:51:30.161974 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/7171eff1-10af-4bab-b43b-02368608ee00-util\") pod \"673e18c374f427d47ebf1299aef3a4ea1aab2302d0e987986af30e26212ztjs\" (UID: \"7171eff1-10af-4bab-b43b-02368608ee00\") " pod="openstack-operators/673e18c374f427d47ebf1299aef3a4ea1aab2302d0e987986af30e26212ztjs" Dec 06 03:51:30 crc kubenswrapper[4980]: I1206 03:51:30.263498 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/7171eff1-10af-4bab-b43b-02368608ee00-bundle\") pod \"673e18c374f427d47ebf1299aef3a4ea1aab2302d0e987986af30e26212ztjs\" (UID: \"7171eff1-10af-4bab-b43b-02368608ee00\") " pod="openstack-operators/673e18c374f427d47ebf1299aef3a4ea1aab2302d0e987986af30e26212ztjs" Dec 06 03:51:30 crc kubenswrapper[4980]: I1206 03:51:30.264129 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/7171eff1-10af-4bab-b43b-02368608ee00-util\") pod \"673e18c374f427d47ebf1299aef3a4ea1aab2302d0e987986af30e26212ztjs\" (UID: \"7171eff1-10af-4bab-b43b-02368608ee00\") " pod="openstack-operators/673e18c374f427d47ebf1299aef3a4ea1aab2302d0e987986af30e26212ztjs" Dec 06 03:51:30 crc kubenswrapper[4980]: I1206 03:51:30.264299 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5qpk4\" (UniqueName: \"kubernetes.io/projected/7171eff1-10af-4bab-b43b-02368608ee00-kube-api-access-5qpk4\") pod \"673e18c374f427d47ebf1299aef3a4ea1aab2302d0e987986af30e26212ztjs\" (UID: \"7171eff1-10af-4bab-b43b-02368608ee00\") " pod="openstack-operators/673e18c374f427d47ebf1299aef3a4ea1aab2302d0e987986af30e26212ztjs" Dec 06 03:51:30 crc kubenswrapper[4980]: I1206 03:51:30.264411 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/7171eff1-10af-4bab-b43b-02368608ee00-bundle\") pod \"673e18c374f427d47ebf1299aef3a4ea1aab2302d0e987986af30e26212ztjs\" (UID: \"7171eff1-10af-4bab-b43b-02368608ee00\") " pod="openstack-operators/673e18c374f427d47ebf1299aef3a4ea1aab2302d0e987986af30e26212ztjs" Dec 06 03:51:30 crc kubenswrapper[4980]: I1206 03:51:30.264900 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/7171eff1-10af-4bab-b43b-02368608ee00-util\") pod \"673e18c374f427d47ebf1299aef3a4ea1aab2302d0e987986af30e26212ztjs\" (UID: \"7171eff1-10af-4bab-b43b-02368608ee00\") " pod="openstack-operators/673e18c374f427d47ebf1299aef3a4ea1aab2302d0e987986af30e26212ztjs" Dec 06 03:51:30 crc kubenswrapper[4980]: I1206 03:51:30.295407 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5qpk4\" (UniqueName: \"kubernetes.io/projected/7171eff1-10af-4bab-b43b-02368608ee00-kube-api-access-5qpk4\") pod \"673e18c374f427d47ebf1299aef3a4ea1aab2302d0e987986af30e26212ztjs\" (UID: \"7171eff1-10af-4bab-b43b-02368608ee00\") " pod="openstack-operators/673e18c374f427d47ebf1299aef3a4ea1aab2302d0e987986af30e26212ztjs" Dec 06 03:51:30 crc kubenswrapper[4980]: I1206 03:51:30.420770 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/673e18c374f427d47ebf1299aef3a4ea1aab2302d0e987986af30e26212ztjs" Dec 06 03:51:30 crc kubenswrapper[4980]: I1206 03:51:30.873301 4980 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/673e18c374f427d47ebf1299aef3a4ea1aab2302d0e987986af30e26212ztjs"] Dec 06 03:51:31 crc kubenswrapper[4980]: I1206 03:51:31.000996 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/673e18c374f427d47ebf1299aef3a4ea1aab2302d0e987986af30e26212ztjs" event={"ID":"7171eff1-10af-4bab-b43b-02368608ee00","Type":"ContainerStarted","Data":"e1ae391d86db9a2fd9e78785dc1e17318f5815df426c4b2f71b8246469382996"} Dec 06 03:51:32 crc kubenswrapper[4980]: I1206 03:51:32.010125 4980 generic.go:334] "Generic (PLEG): container finished" podID="7171eff1-10af-4bab-b43b-02368608ee00" containerID="1c2b88f2043ba0267f8eecae8ecc26fa22d1bb920418f01d64dfd138763bca09" exitCode=0 Dec 06 03:51:32 crc kubenswrapper[4980]: I1206 03:51:32.010178 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/673e18c374f427d47ebf1299aef3a4ea1aab2302d0e987986af30e26212ztjs" event={"ID":"7171eff1-10af-4bab-b43b-02368608ee00","Type":"ContainerDied","Data":"1c2b88f2043ba0267f8eecae8ecc26fa22d1bb920418f01d64dfd138763bca09"} Dec 06 03:51:33 crc kubenswrapper[4980]: I1206 03:51:33.018583 4980 generic.go:334] "Generic (PLEG): container finished" podID="7171eff1-10af-4bab-b43b-02368608ee00" containerID="df9f9a25d6863c4f3447dcc8760c890ce400e1bb62b092300dc1863206ff97ac" exitCode=0 Dec 06 03:51:33 crc kubenswrapper[4980]: I1206 03:51:33.018702 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/673e18c374f427d47ebf1299aef3a4ea1aab2302d0e987986af30e26212ztjs" event={"ID":"7171eff1-10af-4bab-b43b-02368608ee00","Type":"ContainerDied","Data":"df9f9a25d6863c4f3447dcc8760c890ce400e1bb62b092300dc1863206ff97ac"} Dec 06 03:51:34 crc kubenswrapper[4980]: I1206 03:51:34.031885 4980 generic.go:334] "Generic (PLEG): container finished" podID="7171eff1-10af-4bab-b43b-02368608ee00" containerID="dd73dd01ad3e703413e6540aefe327d8d6fca646594169fd39577bd44fc14620" exitCode=0 Dec 06 03:51:34 crc kubenswrapper[4980]: I1206 03:51:34.031935 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/673e18c374f427d47ebf1299aef3a4ea1aab2302d0e987986af30e26212ztjs" event={"ID":"7171eff1-10af-4bab-b43b-02368608ee00","Type":"ContainerDied","Data":"dd73dd01ad3e703413e6540aefe327d8d6fca646594169fd39577bd44fc14620"} Dec 06 03:51:35 crc kubenswrapper[4980]: I1206 03:51:35.305268 4980 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/673e18c374f427d47ebf1299aef3a4ea1aab2302d0e987986af30e26212ztjs" Dec 06 03:51:35 crc kubenswrapper[4980]: I1206 03:51:35.441583 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/7171eff1-10af-4bab-b43b-02368608ee00-util\") pod \"7171eff1-10af-4bab-b43b-02368608ee00\" (UID: \"7171eff1-10af-4bab-b43b-02368608ee00\") " Dec 06 03:51:35 crc kubenswrapper[4980]: I1206 03:51:35.441857 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/7171eff1-10af-4bab-b43b-02368608ee00-bundle\") pod \"7171eff1-10af-4bab-b43b-02368608ee00\" (UID: \"7171eff1-10af-4bab-b43b-02368608ee00\") " Dec 06 03:51:35 crc kubenswrapper[4980]: I1206 03:51:35.441907 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5qpk4\" (UniqueName: \"kubernetes.io/projected/7171eff1-10af-4bab-b43b-02368608ee00-kube-api-access-5qpk4\") pod \"7171eff1-10af-4bab-b43b-02368608ee00\" (UID: \"7171eff1-10af-4bab-b43b-02368608ee00\") " Dec 06 03:51:35 crc kubenswrapper[4980]: I1206 03:51:35.442545 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7171eff1-10af-4bab-b43b-02368608ee00-bundle" (OuterVolumeSpecName: "bundle") pod "7171eff1-10af-4bab-b43b-02368608ee00" (UID: "7171eff1-10af-4bab-b43b-02368608ee00"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 03:51:35 crc kubenswrapper[4980]: I1206 03:51:35.452688 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7171eff1-10af-4bab-b43b-02368608ee00-kube-api-access-5qpk4" (OuterVolumeSpecName: "kube-api-access-5qpk4") pod "7171eff1-10af-4bab-b43b-02368608ee00" (UID: "7171eff1-10af-4bab-b43b-02368608ee00"). InnerVolumeSpecName "kube-api-access-5qpk4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 03:51:35 crc kubenswrapper[4980]: I1206 03:51:35.455781 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7171eff1-10af-4bab-b43b-02368608ee00-util" (OuterVolumeSpecName: "util") pod "7171eff1-10af-4bab-b43b-02368608ee00" (UID: "7171eff1-10af-4bab-b43b-02368608ee00"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 03:51:35 crc kubenswrapper[4980]: I1206 03:51:35.543297 4980 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/7171eff1-10af-4bab-b43b-02368608ee00-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 03:51:35 crc kubenswrapper[4980]: I1206 03:51:35.543338 4980 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5qpk4\" (UniqueName: \"kubernetes.io/projected/7171eff1-10af-4bab-b43b-02368608ee00-kube-api-access-5qpk4\") on node \"crc\" DevicePath \"\"" Dec 06 03:51:35 crc kubenswrapper[4980]: I1206 03:51:35.543347 4980 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/7171eff1-10af-4bab-b43b-02368608ee00-util\") on node \"crc\" DevicePath \"\"" Dec 06 03:51:36 crc kubenswrapper[4980]: I1206 03:51:36.050405 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/rabbitmq-server-0" event={"ID":"92922f24-43cc-445d-a4ad-adcbf68b6636","Type":"ContainerDied","Data":"8d14297fc9e26de59e87c5641efc156518dd1526d321fda49138344216dca6fc"} Dec 06 03:51:36 crc kubenswrapper[4980]: I1206 03:51:36.050342 4980 generic.go:334] "Generic (PLEG): container finished" podID="92922f24-43cc-445d-a4ad-adcbf68b6636" containerID="8d14297fc9e26de59e87c5641efc156518dd1526d321fda49138344216dca6fc" exitCode=0 Dec 06 03:51:36 crc kubenswrapper[4980]: I1206 03:51:36.054835 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/673e18c374f427d47ebf1299aef3a4ea1aab2302d0e987986af30e26212ztjs" event={"ID":"7171eff1-10af-4bab-b43b-02368608ee00","Type":"ContainerDied","Data":"e1ae391d86db9a2fd9e78785dc1e17318f5815df426c4b2f71b8246469382996"} Dec 06 03:51:36 crc kubenswrapper[4980]: I1206 03:51:36.054881 4980 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e1ae391d86db9a2fd9e78785dc1e17318f5815df426c4b2f71b8246469382996" Dec 06 03:51:36 crc kubenswrapper[4980]: I1206 03:51:36.055165 4980 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/673e18c374f427d47ebf1299aef3a4ea1aab2302d0e987986af30e26212ztjs" Dec 06 03:51:37 crc kubenswrapper[4980]: I1206 03:51:37.066576 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/rabbitmq-server-0" event={"ID":"92922f24-43cc-445d-a4ad-adcbf68b6636","Type":"ContainerStarted","Data":"d8e99003864fcca9e18c246f13eb162537692bc504ba074e0b7b349458a7dc79"} Dec 06 03:51:37 crc kubenswrapper[4980]: I1206 03:51:37.067059 4980 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/rabbitmq-server-0" Dec 06 03:51:37 crc kubenswrapper[4980]: I1206 03:51:37.090326 4980 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/rabbitmq-server-0" podStartSLOduration=36.931122013 podStartE2EDuration="44.090290817s" podCreationTimestamp="2025-12-06 03:50:53 +0000 UTC" firstStartedPulling="2025-12-06 03:50:55.311955379 +0000 UTC m=+1014.556706650" lastFinishedPulling="2025-12-06 03:51:02.471124173 +0000 UTC m=+1021.715875454" observedRunningTime="2025-12-06 03:51:37.089142503 +0000 UTC m=+1056.333893794" watchObservedRunningTime="2025-12-06 03:51:37.090290817 +0000 UTC m=+1056.335042088" Dec 06 03:51:46 crc kubenswrapper[4980]: I1206 03:51:46.119552 4980 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/keystone-operator-controller-manager-8485c4fcb7-9qgh9"] Dec 06 03:51:46 crc kubenswrapper[4980]: E1206 03:51:46.120451 4980 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7171eff1-10af-4bab-b43b-02368608ee00" containerName="pull" Dec 06 03:51:46 crc kubenswrapper[4980]: I1206 03:51:46.120468 4980 state_mem.go:107] "Deleted CPUSet assignment" podUID="7171eff1-10af-4bab-b43b-02368608ee00" containerName="pull" Dec 06 03:51:46 crc kubenswrapper[4980]: E1206 03:51:46.120495 4980 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7171eff1-10af-4bab-b43b-02368608ee00" containerName="extract" Dec 06 03:51:46 crc kubenswrapper[4980]: I1206 03:51:46.120504 4980 state_mem.go:107] "Deleted CPUSet assignment" podUID="7171eff1-10af-4bab-b43b-02368608ee00" containerName="extract" Dec 06 03:51:46 crc kubenswrapper[4980]: E1206 03:51:46.120552 4980 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7171eff1-10af-4bab-b43b-02368608ee00" containerName="util" Dec 06 03:51:46 crc kubenswrapper[4980]: I1206 03:51:46.120569 4980 state_mem.go:107] "Deleted CPUSet assignment" podUID="7171eff1-10af-4bab-b43b-02368608ee00" containerName="util" Dec 06 03:51:46 crc kubenswrapper[4980]: I1206 03:51:46.120769 4980 memory_manager.go:354] "RemoveStaleState removing state" podUID="7171eff1-10af-4bab-b43b-02368608ee00" containerName="extract" Dec 06 03:51:46 crc kubenswrapper[4980]: I1206 03:51:46.121754 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-8485c4fcb7-9qgh9" Dec 06 03:51:46 crc kubenswrapper[4980]: I1206 03:51:46.123471 4980 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"keystone-operator-controller-manager-service-cert" Dec 06 03:51:46 crc kubenswrapper[4980]: I1206 03:51:46.124371 4980 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"keystone-operator-controller-manager-dockercfg-vx2wd" Dec 06 03:51:46 crc kubenswrapper[4980]: I1206 03:51:46.163709 4980 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-8485c4fcb7-9qgh9"] Dec 06 03:51:46 crc kubenswrapper[4980]: I1206 03:51:46.308566 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/3f51fcb0-cd67-4d0d-8731-48e926391011-apiservice-cert\") pod \"keystone-operator-controller-manager-8485c4fcb7-9qgh9\" (UID: \"3f51fcb0-cd67-4d0d-8731-48e926391011\") " pod="openstack-operators/keystone-operator-controller-manager-8485c4fcb7-9qgh9" Dec 06 03:51:46 crc kubenswrapper[4980]: I1206 03:51:46.308625 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4z7l7\" (UniqueName: \"kubernetes.io/projected/3f51fcb0-cd67-4d0d-8731-48e926391011-kube-api-access-4z7l7\") pod \"keystone-operator-controller-manager-8485c4fcb7-9qgh9\" (UID: \"3f51fcb0-cd67-4d0d-8731-48e926391011\") " pod="openstack-operators/keystone-operator-controller-manager-8485c4fcb7-9qgh9" Dec 06 03:51:46 crc kubenswrapper[4980]: I1206 03:51:46.308844 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/3f51fcb0-cd67-4d0d-8731-48e926391011-webhook-cert\") pod \"keystone-operator-controller-manager-8485c4fcb7-9qgh9\" (UID: \"3f51fcb0-cd67-4d0d-8731-48e926391011\") " pod="openstack-operators/keystone-operator-controller-manager-8485c4fcb7-9qgh9" Dec 06 03:51:46 crc kubenswrapper[4980]: I1206 03:51:46.409976 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4z7l7\" (UniqueName: \"kubernetes.io/projected/3f51fcb0-cd67-4d0d-8731-48e926391011-kube-api-access-4z7l7\") pod \"keystone-operator-controller-manager-8485c4fcb7-9qgh9\" (UID: \"3f51fcb0-cd67-4d0d-8731-48e926391011\") " pod="openstack-operators/keystone-operator-controller-manager-8485c4fcb7-9qgh9" Dec 06 03:51:46 crc kubenswrapper[4980]: I1206 03:51:46.410126 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/3f51fcb0-cd67-4d0d-8731-48e926391011-webhook-cert\") pod \"keystone-operator-controller-manager-8485c4fcb7-9qgh9\" (UID: \"3f51fcb0-cd67-4d0d-8731-48e926391011\") " pod="openstack-operators/keystone-operator-controller-manager-8485c4fcb7-9qgh9" Dec 06 03:51:46 crc kubenswrapper[4980]: I1206 03:51:46.410280 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/3f51fcb0-cd67-4d0d-8731-48e926391011-apiservice-cert\") pod \"keystone-operator-controller-manager-8485c4fcb7-9qgh9\" (UID: \"3f51fcb0-cd67-4d0d-8731-48e926391011\") " pod="openstack-operators/keystone-operator-controller-manager-8485c4fcb7-9qgh9" Dec 06 03:51:46 crc kubenswrapper[4980]: I1206 03:51:46.416055 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/3f51fcb0-cd67-4d0d-8731-48e926391011-apiservice-cert\") pod \"keystone-operator-controller-manager-8485c4fcb7-9qgh9\" (UID: \"3f51fcb0-cd67-4d0d-8731-48e926391011\") " pod="openstack-operators/keystone-operator-controller-manager-8485c4fcb7-9qgh9" Dec 06 03:51:46 crc kubenswrapper[4980]: I1206 03:51:46.421020 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/3f51fcb0-cd67-4d0d-8731-48e926391011-webhook-cert\") pod \"keystone-operator-controller-manager-8485c4fcb7-9qgh9\" (UID: \"3f51fcb0-cd67-4d0d-8731-48e926391011\") " pod="openstack-operators/keystone-operator-controller-manager-8485c4fcb7-9qgh9" Dec 06 03:51:46 crc kubenswrapper[4980]: I1206 03:51:46.435067 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4z7l7\" (UniqueName: \"kubernetes.io/projected/3f51fcb0-cd67-4d0d-8731-48e926391011-kube-api-access-4z7l7\") pod \"keystone-operator-controller-manager-8485c4fcb7-9qgh9\" (UID: \"3f51fcb0-cd67-4d0d-8731-48e926391011\") " pod="openstack-operators/keystone-operator-controller-manager-8485c4fcb7-9qgh9" Dec 06 03:51:46 crc kubenswrapper[4980]: I1206 03:51:46.463390 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-8485c4fcb7-9qgh9" Dec 06 03:51:47 crc kubenswrapper[4980]: I1206 03:51:47.405479 4980 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-8485c4fcb7-9qgh9"] Dec 06 03:51:47 crc kubenswrapper[4980]: I1206 03:51:47.419141 4980 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 06 03:51:48 crc kubenswrapper[4980]: I1206 03:51:48.154532 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-8485c4fcb7-9qgh9" event={"ID":"3f51fcb0-cd67-4d0d-8731-48e926391011","Type":"ContainerStarted","Data":"139fb89fb90c41ddc53c7073d09fc889c29003c7e04f4bec788a05280bebd012"} Dec 06 03:51:50 crc kubenswrapper[4980]: I1206 03:51:50.170664 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-8485c4fcb7-9qgh9" event={"ID":"3f51fcb0-cd67-4d0d-8731-48e926391011","Type":"ContainerStarted","Data":"36f0dfff3680d0add64557c8a5248f88667b0b83ff1462198958d4818e81a2d7"} Dec 06 03:51:51 crc kubenswrapper[4980]: I1206 03:51:51.178404 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-8485c4fcb7-9qgh9" event={"ID":"3f51fcb0-cd67-4d0d-8731-48e926391011","Type":"ContainerStarted","Data":"d7aaee9a9daacdf7aef993f1eaf3f5635bf0b82feb4c53fbe4fb7715b285b9a6"} Dec 06 03:51:51 crc kubenswrapper[4980]: I1206 03:51:51.178718 4980 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/keystone-operator-controller-manager-8485c4fcb7-9qgh9" Dec 06 03:51:51 crc kubenswrapper[4980]: I1206 03:51:51.202796 4980 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/keystone-operator-controller-manager-8485c4fcb7-9qgh9" podStartSLOduration=2.790263742 podStartE2EDuration="5.202775163s" podCreationTimestamp="2025-12-06 03:51:46 +0000 UTC" firstStartedPulling="2025-12-06 03:51:47.418825628 +0000 UTC m=+1066.663576899" lastFinishedPulling="2025-12-06 03:51:49.831337049 +0000 UTC m=+1069.076088320" observedRunningTime="2025-12-06 03:51:51.198770638 +0000 UTC m=+1070.443521919" watchObservedRunningTime="2025-12-06 03:51:51.202775163 +0000 UTC m=+1070.447526434" Dec 06 03:51:54 crc kubenswrapper[4980]: I1206 03:51:54.830838 4980 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/rabbitmq-server-0" Dec 06 03:51:56 crc kubenswrapper[4980]: I1206 03:51:56.477421 4980 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/keystone-operator-controller-manager-8485c4fcb7-9qgh9" Dec 06 03:52:00 crc kubenswrapper[4980]: I1206 03:52:00.027354 4980 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/keystone-db-create-t9xwz"] Dec 06 03:52:00 crc kubenswrapper[4980]: I1206 03:52:00.028744 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/keystone-db-create-t9xwz" Dec 06 03:52:00 crc kubenswrapper[4980]: I1206 03:52:00.036794 4980 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/keystone-db-create-t9xwz"] Dec 06 03:52:00 crc kubenswrapper[4980]: I1206 03:52:00.112619 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ljjh7\" (UniqueName: \"kubernetes.io/projected/c88434fa-d55f-4ac5-8075-db953773e17c-kube-api-access-ljjh7\") pod \"keystone-db-create-t9xwz\" (UID: \"c88434fa-d55f-4ac5-8075-db953773e17c\") " pod="glance-kuttl-tests/keystone-db-create-t9xwz" Dec 06 03:52:00 crc kubenswrapper[4980]: I1206 03:52:00.214135 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ljjh7\" (UniqueName: \"kubernetes.io/projected/c88434fa-d55f-4ac5-8075-db953773e17c-kube-api-access-ljjh7\") pod \"keystone-db-create-t9xwz\" (UID: \"c88434fa-d55f-4ac5-8075-db953773e17c\") " pod="glance-kuttl-tests/keystone-db-create-t9xwz" Dec 06 03:52:00 crc kubenswrapper[4980]: I1206 03:52:00.244554 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ljjh7\" (UniqueName: \"kubernetes.io/projected/c88434fa-d55f-4ac5-8075-db953773e17c-kube-api-access-ljjh7\") pod \"keystone-db-create-t9xwz\" (UID: \"c88434fa-d55f-4ac5-8075-db953773e17c\") " pod="glance-kuttl-tests/keystone-db-create-t9xwz" Dec 06 03:52:00 crc kubenswrapper[4980]: I1206 03:52:00.358105 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/keystone-db-create-t9xwz" Dec 06 03:52:00 crc kubenswrapper[4980]: I1206 03:52:00.749404 4980 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/keystone-db-create-t9xwz"] Dec 06 03:52:01 crc kubenswrapper[4980]: I1206 03:52:01.241126 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/keystone-db-create-t9xwz" event={"ID":"c88434fa-d55f-4ac5-8075-db953773e17c","Type":"ContainerStarted","Data":"3a35e78496b465762a856b5e824ebddd186c1dfc45adfb5aa4ebd3d1aad84d3f"} Dec 06 03:52:02 crc kubenswrapper[4980]: I1206 03:52:02.254076 4980 generic.go:334] "Generic (PLEG): container finished" podID="c88434fa-d55f-4ac5-8075-db953773e17c" containerID="e6779bc417ecf996156f1751301fa8bddc16f17f41bdc80c112cb5888d28be19" exitCode=0 Dec 06 03:52:02 crc kubenswrapper[4980]: I1206 03:52:02.254228 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/keystone-db-create-t9xwz" event={"ID":"c88434fa-d55f-4ac5-8075-db953773e17c","Type":"ContainerDied","Data":"e6779bc417ecf996156f1751301fa8bddc16f17f41bdc80c112cb5888d28be19"} Dec 06 03:52:02 crc kubenswrapper[4980]: I1206 03:52:02.657638 4980 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/horizon-operator-index-8jmnt"] Dec 06 03:52:02 crc kubenswrapper[4980]: I1206 03:52:02.658725 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-index-8jmnt" Dec 06 03:52:02 crc kubenswrapper[4980]: I1206 03:52:02.661262 4980 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"horizon-operator-index-dockercfg-d8xkm" Dec 06 03:52:02 crc kubenswrapper[4980]: I1206 03:52:02.686286 4980 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-index-8jmnt"] Dec 06 03:52:02 crc kubenswrapper[4980]: I1206 03:52:02.769675 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fbn72\" (UniqueName: \"kubernetes.io/projected/24052d23-b7d3-436b-906f-dcdc363733f9-kube-api-access-fbn72\") pod \"horizon-operator-index-8jmnt\" (UID: \"24052d23-b7d3-436b-906f-dcdc363733f9\") " pod="openstack-operators/horizon-operator-index-8jmnt" Dec 06 03:52:02 crc kubenswrapper[4980]: I1206 03:52:02.871270 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fbn72\" (UniqueName: \"kubernetes.io/projected/24052d23-b7d3-436b-906f-dcdc363733f9-kube-api-access-fbn72\") pod \"horizon-operator-index-8jmnt\" (UID: \"24052d23-b7d3-436b-906f-dcdc363733f9\") " pod="openstack-operators/horizon-operator-index-8jmnt" Dec 06 03:52:02 crc kubenswrapper[4980]: I1206 03:52:02.903382 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fbn72\" (UniqueName: \"kubernetes.io/projected/24052d23-b7d3-436b-906f-dcdc363733f9-kube-api-access-fbn72\") pod \"horizon-operator-index-8jmnt\" (UID: \"24052d23-b7d3-436b-906f-dcdc363733f9\") " pod="openstack-operators/horizon-operator-index-8jmnt" Dec 06 03:52:02 crc kubenswrapper[4980]: I1206 03:52:02.990019 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-index-8jmnt" Dec 06 03:52:03 crc kubenswrapper[4980]: I1206 03:52:03.260145 4980 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-index-8jmnt"] Dec 06 03:52:03 crc kubenswrapper[4980]: I1206 03:52:03.538657 4980 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/keystone-db-create-t9xwz" Dec 06 03:52:03 crc kubenswrapper[4980]: I1206 03:52:03.683202 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ljjh7\" (UniqueName: \"kubernetes.io/projected/c88434fa-d55f-4ac5-8075-db953773e17c-kube-api-access-ljjh7\") pod \"c88434fa-d55f-4ac5-8075-db953773e17c\" (UID: \"c88434fa-d55f-4ac5-8075-db953773e17c\") " Dec 06 03:52:03 crc kubenswrapper[4980]: I1206 03:52:03.688700 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c88434fa-d55f-4ac5-8075-db953773e17c-kube-api-access-ljjh7" (OuterVolumeSpecName: "kube-api-access-ljjh7") pod "c88434fa-d55f-4ac5-8075-db953773e17c" (UID: "c88434fa-d55f-4ac5-8075-db953773e17c"). InnerVolumeSpecName "kube-api-access-ljjh7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 03:52:03 crc kubenswrapper[4980]: I1206 03:52:03.785600 4980 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ljjh7\" (UniqueName: \"kubernetes.io/projected/c88434fa-d55f-4ac5-8075-db953773e17c-kube-api-access-ljjh7\") on node \"crc\" DevicePath \"\"" Dec 06 03:52:04 crc kubenswrapper[4980]: I1206 03:52:04.303328 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-index-8jmnt" event={"ID":"24052d23-b7d3-436b-906f-dcdc363733f9","Type":"ContainerStarted","Data":"c347891296f32460e7ce03a02337ac8e5fb20dc11da7c426acce129be8b17586"} Dec 06 03:52:04 crc kubenswrapper[4980]: I1206 03:52:04.306090 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/keystone-db-create-t9xwz" event={"ID":"c88434fa-d55f-4ac5-8075-db953773e17c","Type":"ContainerDied","Data":"3a35e78496b465762a856b5e824ebddd186c1dfc45adfb5aa4ebd3d1aad84d3f"} Dec 06 03:52:04 crc kubenswrapper[4980]: I1206 03:52:04.306141 4980 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3a35e78496b465762a856b5e824ebddd186c1dfc45adfb5aa4ebd3d1aad84d3f" Dec 06 03:52:04 crc kubenswrapper[4980]: I1206 03:52:04.306158 4980 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/keystone-db-create-t9xwz" Dec 06 03:52:06 crc kubenswrapper[4980]: I1206 03:52:06.254088 4980 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/swift-operator-index-tqvtp"] Dec 06 03:52:06 crc kubenswrapper[4980]: E1206 03:52:06.255369 4980 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c88434fa-d55f-4ac5-8075-db953773e17c" containerName="mariadb-database-create" Dec 06 03:52:06 crc kubenswrapper[4980]: I1206 03:52:06.255448 4980 state_mem.go:107] "Deleted CPUSet assignment" podUID="c88434fa-d55f-4ac5-8075-db953773e17c" containerName="mariadb-database-create" Dec 06 03:52:06 crc kubenswrapper[4980]: I1206 03:52:06.256139 4980 memory_manager.go:354] "RemoveStaleState removing state" podUID="c88434fa-d55f-4ac5-8075-db953773e17c" containerName="mariadb-database-create" Dec 06 03:52:06 crc kubenswrapper[4980]: I1206 03:52:06.257379 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-index-tqvtp" Dec 06 03:52:06 crc kubenswrapper[4980]: I1206 03:52:06.261416 4980 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"swift-operator-index-dockercfg-d27tb" Dec 06 03:52:06 crc kubenswrapper[4980]: I1206 03:52:06.265449 4980 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-index-tqvtp"] Dec 06 03:52:06 crc kubenswrapper[4980]: I1206 03:52:06.337619 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8lngc\" (UniqueName: \"kubernetes.io/projected/ece95a8d-b36b-425e-9b6f-6ee1157e9e98-kube-api-access-8lngc\") pod \"swift-operator-index-tqvtp\" (UID: \"ece95a8d-b36b-425e-9b6f-6ee1157e9e98\") " pod="openstack-operators/swift-operator-index-tqvtp" Dec 06 03:52:06 crc kubenswrapper[4980]: I1206 03:52:06.439508 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8lngc\" (UniqueName: \"kubernetes.io/projected/ece95a8d-b36b-425e-9b6f-6ee1157e9e98-kube-api-access-8lngc\") pod \"swift-operator-index-tqvtp\" (UID: \"ece95a8d-b36b-425e-9b6f-6ee1157e9e98\") " pod="openstack-operators/swift-operator-index-tqvtp" Dec 06 03:52:06 crc kubenswrapper[4980]: I1206 03:52:06.458363 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8lngc\" (UniqueName: \"kubernetes.io/projected/ece95a8d-b36b-425e-9b6f-6ee1157e9e98-kube-api-access-8lngc\") pod \"swift-operator-index-tqvtp\" (UID: \"ece95a8d-b36b-425e-9b6f-6ee1157e9e98\") " pod="openstack-operators/swift-operator-index-tqvtp" Dec 06 03:52:06 crc kubenswrapper[4980]: I1206 03:52:06.583855 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-index-tqvtp" Dec 06 03:52:07 crc kubenswrapper[4980]: I1206 03:52:07.151116 4980 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-index-tqvtp"] Dec 06 03:52:07 crc kubenswrapper[4980]: I1206 03:52:07.332447 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-index-8jmnt" event={"ID":"24052d23-b7d3-436b-906f-dcdc363733f9","Type":"ContainerStarted","Data":"c986b906e838e095a11e10c3938d8fde15c6e1f4bd00fe96e731fd512068de38"} Dec 06 03:52:07 crc kubenswrapper[4980]: I1206 03:52:07.333535 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-index-tqvtp" event={"ID":"ece95a8d-b36b-425e-9b6f-6ee1157e9e98","Type":"ContainerStarted","Data":"bd273b0e7d7640b9f539fae42bbaaf8e2e3f674d5f40b109a33c8b35951037a2"} Dec 06 03:52:07 crc kubenswrapper[4980]: I1206 03:52:07.351876 4980 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/horizon-operator-index-8jmnt" podStartSLOduration=2.27525578 podStartE2EDuration="5.35184048s" podCreationTimestamp="2025-12-06 03:52:02 +0000 UTC" firstStartedPulling="2025-12-06 03:52:03.275310837 +0000 UTC m=+1082.520062118" lastFinishedPulling="2025-12-06 03:52:06.351895537 +0000 UTC m=+1085.596646818" observedRunningTime="2025-12-06 03:52:07.348873125 +0000 UTC m=+1086.593624396" watchObservedRunningTime="2025-12-06 03:52:07.35184048 +0000 UTC m=+1086.596591741" Dec 06 03:52:09 crc kubenswrapper[4980]: I1206 03:52:09.349577 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-index-tqvtp" event={"ID":"ece95a8d-b36b-425e-9b6f-6ee1157e9e98","Type":"ContainerStarted","Data":"803ca4271887ee9ea6fc64ec857d46d6bbabe5107c579622ad0ad44fa7bc966c"} Dec 06 03:52:09 crc kubenswrapper[4980]: I1206 03:52:09.363244 4980 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/swift-operator-index-tqvtp" podStartSLOduration=1.374723109 podStartE2EDuration="3.363222627s" podCreationTimestamp="2025-12-06 03:52:06 +0000 UTC" firstStartedPulling="2025-12-06 03:52:07.1560364 +0000 UTC m=+1086.400787671" lastFinishedPulling="2025-12-06 03:52:09.144535898 +0000 UTC m=+1088.389287189" observedRunningTime="2025-12-06 03:52:09.362118025 +0000 UTC m=+1088.606869336" watchObservedRunningTime="2025-12-06 03:52:09.363222627 +0000 UTC m=+1088.607973898" Dec 06 03:52:09 crc kubenswrapper[4980]: I1206 03:52:09.935087 4980 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/keystone-0b06-account-create-pf8hh"] Dec 06 03:52:09 crc kubenswrapper[4980]: I1206 03:52:09.936462 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/keystone-0b06-account-create-pf8hh" Dec 06 03:52:09 crc kubenswrapper[4980]: I1206 03:52:09.939387 4980 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"keystone-db-secret" Dec 06 03:52:09 crc kubenswrapper[4980]: I1206 03:52:09.945014 4980 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/keystone-0b06-account-create-pf8hh"] Dec 06 03:52:09 crc kubenswrapper[4980]: I1206 03:52:09.989803 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ks5vq\" (UniqueName: \"kubernetes.io/projected/f76dd592-3194-4b0a-81e2-fc87f32f3214-kube-api-access-ks5vq\") pod \"keystone-0b06-account-create-pf8hh\" (UID: \"f76dd592-3194-4b0a-81e2-fc87f32f3214\") " pod="glance-kuttl-tests/keystone-0b06-account-create-pf8hh" Dec 06 03:52:10 crc kubenswrapper[4980]: I1206 03:52:10.091553 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ks5vq\" (UniqueName: \"kubernetes.io/projected/f76dd592-3194-4b0a-81e2-fc87f32f3214-kube-api-access-ks5vq\") pod \"keystone-0b06-account-create-pf8hh\" (UID: \"f76dd592-3194-4b0a-81e2-fc87f32f3214\") " pod="glance-kuttl-tests/keystone-0b06-account-create-pf8hh" Dec 06 03:52:10 crc kubenswrapper[4980]: I1206 03:52:10.113575 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ks5vq\" (UniqueName: \"kubernetes.io/projected/f76dd592-3194-4b0a-81e2-fc87f32f3214-kube-api-access-ks5vq\") pod \"keystone-0b06-account-create-pf8hh\" (UID: \"f76dd592-3194-4b0a-81e2-fc87f32f3214\") " pod="glance-kuttl-tests/keystone-0b06-account-create-pf8hh" Dec 06 03:52:10 crc kubenswrapper[4980]: I1206 03:52:10.256922 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/keystone-0b06-account-create-pf8hh" Dec 06 03:52:10 crc kubenswrapper[4980]: I1206 03:52:10.670928 4980 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/keystone-0b06-account-create-pf8hh"] Dec 06 03:52:11 crc kubenswrapper[4980]: I1206 03:52:11.368334 4980 generic.go:334] "Generic (PLEG): container finished" podID="f76dd592-3194-4b0a-81e2-fc87f32f3214" containerID="c451c6ce6fb1fbc8b8c12ae524173dd2515e9eb63d5a7cea60858425ae32143a" exitCode=0 Dec 06 03:52:11 crc kubenswrapper[4980]: I1206 03:52:11.368428 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/keystone-0b06-account-create-pf8hh" event={"ID":"f76dd592-3194-4b0a-81e2-fc87f32f3214","Type":"ContainerDied","Data":"c451c6ce6fb1fbc8b8c12ae524173dd2515e9eb63d5a7cea60858425ae32143a"} Dec 06 03:52:11 crc kubenswrapper[4980]: I1206 03:52:11.368543 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/keystone-0b06-account-create-pf8hh" event={"ID":"f76dd592-3194-4b0a-81e2-fc87f32f3214","Type":"ContainerStarted","Data":"d49a8b1d663c2b4949874d0483f85af6966565048387203c3799e09a4a77ab4c"} Dec 06 03:52:12 crc kubenswrapper[4980]: I1206 03:52:12.237910 4980 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/swift-operator-index-tqvtp"] Dec 06 03:52:12 crc kubenswrapper[4980]: I1206 03:52:12.239251 4980 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/swift-operator-index-tqvtp" podUID="ece95a8d-b36b-425e-9b6f-6ee1157e9e98" containerName="registry-server" containerID="cri-o://803ca4271887ee9ea6fc64ec857d46d6bbabe5107c579622ad0ad44fa7bc966c" gracePeriod=2 Dec 06 03:52:12 crc kubenswrapper[4980]: I1206 03:52:12.640857 4980 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/keystone-0b06-account-create-pf8hh" Dec 06 03:52:12 crc kubenswrapper[4980]: I1206 03:52:12.728985 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ks5vq\" (UniqueName: \"kubernetes.io/projected/f76dd592-3194-4b0a-81e2-fc87f32f3214-kube-api-access-ks5vq\") pod \"f76dd592-3194-4b0a-81e2-fc87f32f3214\" (UID: \"f76dd592-3194-4b0a-81e2-fc87f32f3214\") " Dec 06 03:52:12 crc kubenswrapper[4980]: I1206 03:52:12.735912 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f76dd592-3194-4b0a-81e2-fc87f32f3214-kube-api-access-ks5vq" (OuterVolumeSpecName: "kube-api-access-ks5vq") pod "f76dd592-3194-4b0a-81e2-fc87f32f3214" (UID: "f76dd592-3194-4b0a-81e2-fc87f32f3214"). InnerVolumeSpecName "kube-api-access-ks5vq". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 03:52:12 crc kubenswrapper[4980]: I1206 03:52:12.832504 4980 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ks5vq\" (UniqueName: \"kubernetes.io/projected/f76dd592-3194-4b0a-81e2-fc87f32f3214-kube-api-access-ks5vq\") on node \"crc\" DevicePath \"\"" Dec 06 03:52:12 crc kubenswrapper[4980]: I1206 03:52:12.850799 4980 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/swift-operator-index-rxptj"] Dec 06 03:52:12 crc kubenswrapper[4980]: E1206 03:52:12.851126 4980 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f76dd592-3194-4b0a-81e2-fc87f32f3214" containerName="mariadb-account-create" Dec 06 03:52:12 crc kubenswrapper[4980]: I1206 03:52:12.851143 4980 state_mem.go:107] "Deleted CPUSet assignment" podUID="f76dd592-3194-4b0a-81e2-fc87f32f3214" containerName="mariadb-account-create" Dec 06 03:52:12 crc kubenswrapper[4980]: I1206 03:52:12.851257 4980 memory_manager.go:354] "RemoveStaleState removing state" podUID="f76dd592-3194-4b0a-81e2-fc87f32f3214" containerName="mariadb-account-create" Dec 06 03:52:12 crc kubenswrapper[4980]: I1206 03:52:12.851977 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-index-rxptj" Dec 06 03:52:12 crc kubenswrapper[4980]: I1206 03:52:12.873018 4980 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-index-rxptj"] Dec 06 03:52:12 crc kubenswrapper[4980]: I1206 03:52:12.934386 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m6r44\" (UniqueName: \"kubernetes.io/projected/713b6c54-f0c9-4c60-9f98-0c2e57dae276-kube-api-access-m6r44\") pod \"swift-operator-index-rxptj\" (UID: \"713b6c54-f0c9-4c60-9f98-0c2e57dae276\") " pod="openstack-operators/swift-operator-index-rxptj" Dec 06 03:52:12 crc kubenswrapper[4980]: I1206 03:52:12.991034 4980 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/horizon-operator-index-8jmnt" Dec 06 03:52:12 crc kubenswrapper[4980]: I1206 03:52:12.991087 4980 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/horizon-operator-index-8jmnt" Dec 06 03:52:13 crc kubenswrapper[4980]: I1206 03:52:13.027238 4980 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/horizon-operator-index-8jmnt" Dec 06 03:52:13 crc kubenswrapper[4980]: I1206 03:52:13.035454 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m6r44\" (UniqueName: \"kubernetes.io/projected/713b6c54-f0c9-4c60-9f98-0c2e57dae276-kube-api-access-m6r44\") pod \"swift-operator-index-rxptj\" (UID: \"713b6c54-f0c9-4c60-9f98-0c2e57dae276\") " pod="openstack-operators/swift-operator-index-rxptj" Dec 06 03:52:13 crc kubenswrapper[4980]: I1206 03:52:13.063636 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m6r44\" (UniqueName: \"kubernetes.io/projected/713b6c54-f0c9-4c60-9f98-0c2e57dae276-kube-api-access-m6r44\") pod \"swift-operator-index-rxptj\" (UID: \"713b6c54-f0c9-4c60-9f98-0c2e57dae276\") " pod="openstack-operators/swift-operator-index-rxptj" Dec 06 03:52:13 crc kubenswrapper[4980]: I1206 03:52:13.151997 4980 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-index-tqvtp" Dec 06 03:52:13 crc kubenswrapper[4980]: I1206 03:52:13.182571 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-index-rxptj" Dec 06 03:52:13 crc kubenswrapper[4980]: I1206 03:52:13.237925 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8lngc\" (UniqueName: \"kubernetes.io/projected/ece95a8d-b36b-425e-9b6f-6ee1157e9e98-kube-api-access-8lngc\") pod \"ece95a8d-b36b-425e-9b6f-6ee1157e9e98\" (UID: \"ece95a8d-b36b-425e-9b6f-6ee1157e9e98\") " Dec 06 03:52:13 crc kubenswrapper[4980]: I1206 03:52:13.242771 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ece95a8d-b36b-425e-9b6f-6ee1157e9e98-kube-api-access-8lngc" (OuterVolumeSpecName: "kube-api-access-8lngc") pod "ece95a8d-b36b-425e-9b6f-6ee1157e9e98" (UID: "ece95a8d-b36b-425e-9b6f-6ee1157e9e98"). InnerVolumeSpecName "kube-api-access-8lngc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 03:52:13 crc kubenswrapper[4980]: I1206 03:52:13.339402 4980 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8lngc\" (UniqueName: \"kubernetes.io/projected/ece95a8d-b36b-425e-9b6f-6ee1157e9e98-kube-api-access-8lngc\") on node \"crc\" DevicePath \"\"" Dec 06 03:52:13 crc kubenswrapper[4980]: I1206 03:52:13.386394 4980 generic.go:334] "Generic (PLEG): container finished" podID="ece95a8d-b36b-425e-9b6f-6ee1157e9e98" containerID="803ca4271887ee9ea6fc64ec857d46d6bbabe5107c579622ad0ad44fa7bc966c" exitCode=0 Dec 06 03:52:13 crc kubenswrapper[4980]: I1206 03:52:13.386473 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-index-tqvtp" event={"ID":"ece95a8d-b36b-425e-9b6f-6ee1157e9e98","Type":"ContainerDied","Data":"803ca4271887ee9ea6fc64ec857d46d6bbabe5107c579622ad0ad44fa7bc966c"} Dec 06 03:52:13 crc kubenswrapper[4980]: I1206 03:52:13.387227 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-index-tqvtp" event={"ID":"ece95a8d-b36b-425e-9b6f-6ee1157e9e98","Type":"ContainerDied","Data":"bd273b0e7d7640b9f539fae42bbaaf8e2e3f674d5f40b109a33c8b35951037a2"} Dec 06 03:52:13 crc kubenswrapper[4980]: I1206 03:52:13.387554 4980 scope.go:117] "RemoveContainer" containerID="803ca4271887ee9ea6fc64ec857d46d6bbabe5107c579622ad0ad44fa7bc966c" Dec 06 03:52:13 crc kubenswrapper[4980]: I1206 03:52:13.386576 4980 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-index-tqvtp" Dec 06 03:52:13 crc kubenswrapper[4980]: I1206 03:52:13.395363 4980 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/keystone-0b06-account-create-pf8hh" Dec 06 03:52:13 crc kubenswrapper[4980]: I1206 03:52:13.395914 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/keystone-0b06-account-create-pf8hh" event={"ID":"f76dd592-3194-4b0a-81e2-fc87f32f3214","Type":"ContainerDied","Data":"d49a8b1d663c2b4949874d0483f85af6966565048387203c3799e09a4a77ab4c"} Dec 06 03:52:13 crc kubenswrapper[4980]: I1206 03:52:13.395942 4980 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d49a8b1d663c2b4949874d0483f85af6966565048387203c3799e09a4a77ab4c" Dec 06 03:52:13 crc kubenswrapper[4980]: I1206 03:52:13.408620 4980 scope.go:117] "RemoveContainer" containerID="803ca4271887ee9ea6fc64ec857d46d6bbabe5107c579622ad0ad44fa7bc966c" Dec 06 03:52:13 crc kubenswrapper[4980]: E1206 03:52:13.409634 4980 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"803ca4271887ee9ea6fc64ec857d46d6bbabe5107c579622ad0ad44fa7bc966c\": container with ID starting with 803ca4271887ee9ea6fc64ec857d46d6bbabe5107c579622ad0ad44fa7bc966c not found: ID does not exist" containerID="803ca4271887ee9ea6fc64ec857d46d6bbabe5107c579622ad0ad44fa7bc966c" Dec 06 03:52:13 crc kubenswrapper[4980]: I1206 03:52:13.409680 4980 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"803ca4271887ee9ea6fc64ec857d46d6bbabe5107c579622ad0ad44fa7bc966c"} err="failed to get container status \"803ca4271887ee9ea6fc64ec857d46d6bbabe5107c579622ad0ad44fa7bc966c\": rpc error: code = NotFound desc = could not find container \"803ca4271887ee9ea6fc64ec857d46d6bbabe5107c579622ad0ad44fa7bc966c\": container with ID starting with 803ca4271887ee9ea6fc64ec857d46d6bbabe5107c579622ad0ad44fa7bc966c not found: ID does not exist" Dec 06 03:52:13 crc kubenswrapper[4980]: I1206 03:52:13.424722 4980 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/swift-operator-index-tqvtp"] Dec 06 03:52:13 crc kubenswrapper[4980]: I1206 03:52:13.427979 4980 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/swift-operator-index-tqvtp"] Dec 06 03:52:13 crc kubenswrapper[4980]: I1206 03:52:13.442144 4980 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/horizon-operator-index-8jmnt" Dec 06 03:52:13 crc kubenswrapper[4980]: I1206 03:52:13.584805 4980 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-index-rxptj"] Dec 06 03:52:14 crc kubenswrapper[4980]: I1206 03:52:14.428643 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-index-rxptj" event={"ID":"713b6c54-f0c9-4c60-9f98-0c2e57dae276","Type":"ContainerStarted","Data":"d48e0f1fc7fcd1432176c4678f1d2a3c6bd9b5bbb233ba0f1e16376804702844"} Dec 06 03:52:15 crc kubenswrapper[4980]: I1206 03:52:15.192798 4980 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ece95a8d-b36b-425e-9b6f-6ee1157e9e98" path="/var/lib/kubelet/pods/ece95a8d-b36b-425e-9b6f-6ee1157e9e98/volumes" Dec 06 03:52:15 crc kubenswrapper[4980]: I1206 03:52:15.425410 4980 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/keystone-db-sync-5bvrm"] Dec 06 03:52:15 crc kubenswrapper[4980]: E1206 03:52:15.425796 4980 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ece95a8d-b36b-425e-9b6f-6ee1157e9e98" containerName="registry-server" Dec 06 03:52:15 crc kubenswrapper[4980]: I1206 03:52:15.425824 4980 state_mem.go:107] "Deleted CPUSet assignment" podUID="ece95a8d-b36b-425e-9b6f-6ee1157e9e98" containerName="registry-server" Dec 06 03:52:15 crc kubenswrapper[4980]: I1206 03:52:15.426069 4980 memory_manager.go:354] "RemoveStaleState removing state" podUID="ece95a8d-b36b-425e-9b6f-6ee1157e9e98" containerName="registry-server" Dec 06 03:52:15 crc kubenswrapper[4980]: I1206 03:52:15.426725 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/keystone-db-sync-5bvrm" Dec 06 03:52:15 crc kubenswrapper[4980]: I1206 03:52:15.428941 4980 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"keystone-keystone-dockercfg-hsvtv" Dec 06 03:52:15 crc kubenswrapper[4980]: I1206 03:52:15.429099 4980 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"keystone-config-data" Dec 06 03:52:15 crc kubenswrapper[4980]: I1206 03:52:15.429245 4980 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"keystone-scripts" Dec 06 03:52:15 crc kubenswrapper[4980]: I1206 03:52:15.436908 4980 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"keystone" Dec 06 03:52:15 crc kubenswrapper[4980]: I1206 03:52:15.470567 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d8d8646d-f06f-403c-9ed7-e870564c7329-config-data\") pod \"keystone-db-sync-5bvrm\" (UID: \"d8d8646d-f06f-403c-9ed7-e870564c7329\") " pod="glance-kuttl-tests/keystone-db-sync-5bvrm" Dec 06 03:52:15 crc kubenswrapper[4980]: I1206 03:52:15.474271 4980 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/keystone-db-sync-5bvrm"] Dec 06 03:52:15 crc kubenswrapper[4980]: I1206 03:52:15.479537 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-index-rxptj" event={"ID":"713b6c54-f0c9-4c60-9f98-0c2e57dae276","Type":"ContainerStarted","Data":"14c03c57b0ef4bd49cc1802f62ca631ae7a9b8ce5863fa52e069a0e118a9092c"} Dec 06 03:52:15 crc kubenswrapper[4980]: I1206 03:52:15.501593 4980 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/swift-operator-index-rxptj" podStartSLOduration=2.562577483 podStartE2EDuration="3.501575191s" podCreationTimestamp="2025-12-06 03:52:12 +0000 UTC" firstStartedPulling="2025-12-06 03:52:13.594311852 +0000 UTC m=+1092.839063133" lastFinishedPulling="2025-12-06 03:52:14.53330957 +0000 UTC m=+1093.778060841" observedRunningTime="2025-12-06 03:52:15.499365467 +0000 UTC m=+1094.744116748" watchObservedRunningTime="2025-12-06 03:52:15.501575191 +0000 UTC m=+1094.746326462" Dec 06 03:52:15 crc kubenswrapper[4980]: I1206 03:52:15.572429 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fmpvp\" (UniqueName: \"kubernetes.io/projected/d8d8646d-f06f-403c-9ed7-e870564c7329-kube-api-access-fmpvp\") pod \"keystone-db-sync-5bvrm\" (UID: \"d8d8646d-f06f-403c-9ed7-e870564c7329\") " pod="glance-kuttl-tests/keystone-db-sync-5bvrm" Dec 06 03:52:15 crc kubenswrapper[4980]: I1206 03:52:15.572816 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d8d8646d-f06f-403c-9ed7-e870564c7329-config-data\") pod \"keystone-db-sync-5bvrm\" (UID: \"d8d8646d-f06f-403c-9ed7-e870564c7329\") " pod="glance-kuttl-tests/keystone-db-sync-5bvrm" Dec 06 03:52:15 crc kubenswrapper[4980]: I1206 03:52:15.583398 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d8d8646d-f06f-403c-9ed7-e870564c7329-config-data\") pod \"keystone-db-sync-5bvrm\" (UID: \"d8d8646d-f06f-403c-9ed7-e870564c7329\") " pod="glance-kuttl-tests/keystone-db-sync-5bvrm" Dec 06 03:52:15 crc kubenswrapper[4980]: I1206 03:52:15.673796 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fmpvp\" (UniqueName: \"kubernetes.io/projected/d8d8646d-f06f-403c-9ed7-e870564c7329-kube-api-access-fmpvp\") pod \"keystone-db-sync-5bvrm\" (UID: \"d8d8646d-f06f-403c-9ed7-e870564c7329\") " pod="glance-kuttl-tests/keystone-db-sync-5bvrm" Dec 06 03:52:15 crc kubenswrapper[4980]: I1206 03:52:15.692958 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fmpvp\" (UniqueName: \"kubernetes.io/projected/d8d8646d-f06f-403c-9ed7-e870564c7329-kube-api-access-fmpvp\") pod \"keystone-db-sync-5bvrm\" (UID: \"d8d8646d-f06f-403c-9ed7-e870564c7329\") " pod="glance-kuttl-tests/keystone-db-sync-5bvrm" Dec 06 03:52:15 crc kubenswrapper[4980]: I1206 03:52:15.759726 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/keystone-db-sync-5bvrm" Dec 06 03:52:16 crc kubenswrapper[4980]: I1206 03:52:16.189628 4980 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/keystone-db-sync-5bvrm"] Dec 06 03:52:16 crc kubenswrapper[4980]: I1206 03:52:16.489128 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/keystone-db-sync-5bvrm" event={"ID":"d8d8646d-f06f-403c-9ed7-e870564c7329","Type":"ContainerStarted","Data":"aa77285cd65a07f72c6cd39d205e9db53c986d8a6344e85cc643379efedd8cb6"} Dec 06 03:52:23 crc kubenswrapper[4980]: I1206 03:52:23.183820 4980 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/swift-operator-index-rxptj" Dec 06 03:52:23 crc kubenswrapper[4980]: I1206 03:52:23.184493 4980 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/swift-operator-index-rxptj" Dec 06 03:52:23 crc kubenswrapper[4980]: I1206 03:52:23.213057 4980 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/swift-operator-index-rxptj" Dec 06 03:52:23 crc kubenswrapper[4980]: I1206 03:52:23.564162 4980 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/swift-operator-index-rxptj" Dec 06 03:52:25 crc kubenswrapper[4980]: I1206 03:52:25.137161 4980 patch_prober.go:28] interesting pod/machine-config-daemon-r5zfb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 03:52:25 crc kubenswrapper[4980]: I1206 03:52:25.137575 4980 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-r5zfb" podUID="320f44d1-a671-4a91-b328-a8b0fdd8f23a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 03:52:25 crc kubenswrapper[4980]: I1206 03:52:25.552302 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/keystone-db-sync-5bvrm" event={"ID":"d8d8646d-f06f-403c-9ed7-e870564c7329","Type":"ContainerStarted","Data":"3e4480c29db3d34e5961de14acf7a1876fe97eb74b5a64c421b4e09e932d21ae"} Dec 06 03:52:25 crc kubenswrapper[4980]: I1206 03:52:25.584976 4980 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/keystone-db-sync-5bvrm" podStartSLOduration=1.538912321 podStartE2EDuration="10.584923768s" podCreationTimestamp="2025-12-06 03:52:15 +0000 UTC" firstStartedPulling="2025-12-06 03:52:16.20005104 +0000 UTC m=+1095.444802311" lastFinishedPulling="2025-12-06 03:52:25.246062487 +0000 UTC m=+1104.490813758" observedRunningTime="2025-12-06 03:52:25.574440176 +0000 UTC m=+1104.819191457" watchObservedRunningTime="2025-12-06 03:52:25.584923768 +0000 UTC m=+1104.829675089" Dec 06 03:52:27 crc kubenswrapper[4980]: I1206 03:52:27.088040 4980 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/c03115aeb6cae4480b1000a58090a522fd8798c774d18bf66e63c4d3d15vssx"] Dec 06 03:52:27 crc kubenswrapper[4980]: I1206 03:52:27.089592 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/c03115aeb6cae4480b1000a58090a522fd8798c774d18bf66e63c4d3d15vssx" Dec 06 03:52:27 crc kubenswrapper[4980]: I1206 03:52:27.091206 4980 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"default-dockercfg-brs8h" Dec 06 03:52:27 crc kubenswrapper[4980]: I1206 03:52:27.100015 4980 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/c03115aeb6cae4480b1000a58090a522fd8798c774d18bf66e63c4d3d15vssx"] Dec 06 03:52:27 crc kubenswrapper[4980]: I1206 03:52:27.108536 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mp6d8\" (UniqueName: \"kubernetes.io/projected/b59d13fa-5f2a-4f67-9502-4b81c035ee10-kube-api-access-mp6d8\") pod \"c03115aeb6cae4480b1000a58090a522fd8798c774d18bf66e63c4d3d15vssx\" (UID: \"b59d13fa-5f2a-4f67-9502-4b81c035ee10\") " pod="openstack-operators/c03115aeb6cae4480b1000a58090a522fd8798c774d18bf66e63c4d3d15vssx" Dec 06 03:52:27 crc kubenswrapper[4980]: I1206 03:52:27.108608 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/b59d13fa-5f2a-4f67-9502-4b81c035ee10-bundle\") pod \"c03115aeb6cae4480b1000a58090a522fd8798c774d18bf66e63c4d3d15vssx\" (UID: \"b59d13fa-5f2a-4f67-9502-4b81c035ee10\") " pod="openstack-operators/c03115aeb6cae4480b1000a58090a522fd8798c774d18bf66e63c4d3d15vssx" Dec 06 03:52:27 crc kubenswrapper[4980]: I1206 03:52:27.108652 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/b59d13fa-5f2a-4f67-9502-4b81c035ee10-util\") pod \"c03115aeb6cae4480b1000a58090a522fd8798c774d18bf66e63c4d3d15vssx\" (UID: \"b59d13fa-5f2a-4f67-9502-4b81c035ee10\") " pod="openstack-operators/c03115aeb6cae4480b1000a58090a522fd8798c774d18bf66e63c4d3d15vssx" Dec 06 03:52:27 crc kubenswrapper[4980]: I1206 03:52:27.210278 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mp6d8\" (UniqueName: \"kubernetes.io/projected/b59d13fa-5f2a-4f67-9502-4b81c035ee10-kube-api-access-mp6d8\") pod \"c03115aeb6cae4480b1000a58090a522fd8798c774d18bf66e63c4d3d15vssx\" (UID: \"b59d13fa-5f2a-4f67-9502-4b81c035ee10\") " pod="openstack-operators/c03115aeb6cae4480b1000a58090a522fd8798c774d18bf66e63c4d3d15vssx" Dec 06 03:52:27 crc kubenswrapper[4980]: I1206 03:52:27.210401 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/b59d13fa-5f2a-4f67-9502-4b81c035ee10-bundle\") pod \"c03115aeb6cae4480b1000a58090a522fd8798c774d18bf66e63c4d3d15vssx\" (UID: \"b59d13fa-5f2a-4f67-9502-4b81c035ee10\") " pod="openstack-operators/c03115aeb6cae4480b1000a58090a522fd8798c774d18bf66e63c4d3d15vssx" Dec 06 03:52:27 crc kubenswrapper[4980]: I1206 03:52:27.210465 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/b59d13fa-5f2a-4f67-9502-4b81c035ee10-util\") pod \"c03115aeb6cae4480b1000a58090a522fd8798c774d18bf66e63c4d3d15vssx\" (UID: \"b59d13fa-5f2a-4f67-9502-4b81c035ee10\") " pod="openstack-operators/c03115aeb6cae4480b1000a58090a522fd8798c774d18bf66e63c4d3d15vssx" Dec 06 03:52:27 crc kubenswrapper[4980]: I1206 03:52:27.211092 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/b59d13fa-5f2a-4f67-9502-4b81c035ee10-util\") pod \"c03115aeb6cae4480b1000a58090a522fd8798c774d18bf66e63c4d3d15vssx\" (UID: \"b59d13fa-5f2a-4f67-9502-4b81c035ee10\") " pod="openstack-operators/c03115aeb6cae4480b1000a58090a522fd8798c774d18bf66e63c4d3d15vssx" Dec 06 03:52:27 crc kubenswrapper[4980]: I1206 03:52:27.211737 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/b59d13fa-5f2a-4f67-9502-4b81c035ee10-bundle\") pod \"c03115aeb6cae4480b1000a58090a522fd8798c774d18bf66e63c4d3d15vssx\" (UID: \"b59d13fa-5f2a-4f67-9502-4b81c035ee10\") " pod="openstack-operators/c03115aeb6cae4480b1000a58090a522fd8798c774d18bf66e63c4d3d15vssx" Dec 06 03:52:27 crc kubenswrapper[4980]: I1206 03:52:27.233068 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mp6d8\" (UniqueName: \"kubernetes.io/projected/b59d13fa-5f2a-4f67-9502-4b81c035ee10-kube-api-access-mp6d8\") pod \"c03115aeb6cae4480b1000a58090a522fd8798c774d18bf66e63c4d3d15vssx\" (UID: \"b59d13fa-5f2a-4f67-9502-4b81c035ee10\") " pod="openstack-operators/c03115aeb6cae4480b1000a58090a522fd8798c774d18bf66e63c4d3d15vssx" Dec 06 03:52:27 crc kubenswrapper[4980]: I1206 03:52:27.409273 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/c03115aeb6cae4480b1000a58090a522fd8798c774d18bf66e63c4d3d15vssx" Dec 06 03:52:27 crc kubenswrapper[4980]: I1206 03:52:27.828338 4980 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/c03115aeb6cae4480b1000a58090a522fd8798c774d18bf66e63c4d3d15vssx"] Dec 06 03:52:28 crc kubenswrapper[4980]: I1206 03:52:28.078407 4980 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/4d9ed08db2179c6d86e46659534222f77f18c6c6d13f1f155004bdcf89l5n8z"] Dec 06 03:52:28 crc kubenswrapper[4980]: I1206 03:52:28.080077 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/4d9ed08db2179c6d86e46659534222f77f18c6c6d13f1f155004bdcf89l5n8z" Dec 06 03:52:28 crc kubenswrapper[4980]: I1206 03:52:28.093703 4980 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/4d9ed08db2179c6d86e46659534222f77f18c6c6d13f1f155004bdcf89l5n8z"] Dec 06 03:52:28 crc kubenswrapper[4980]: I1206 03:52:28.121521 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4hx4m\" (UniqueName: \"kubernetes.io/projected/6674a828-934e-4b14-98be-740dad507b95-kube-api-access-4hx4m\") pod \"4d9ed08db2179c6d86e46659534222f77f18c6c6d13f1f155004bdcf89l5n8z\" (UID: \"6674a828-934e-4b14-98be-740dad507b95\") " pod="openstack-operators/4d9ed08db2179c6d86e46659534222f77f18c6c6d13f1f155004bdcf89l5n8z" Dec 06 03:52:28 crc kubenswrapper[4980]: I1206 03:52:28.121689 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/6674a828-934e-4b14-98be-740dad507b95-bundle\") pod \"4d9ed08db2179c6d86e46659534222f77f18c6c6d13f1f155004bdcf89l5n8z\" (UID: \"6674a828-934e-4b14-98be-740dad507b95\") " pod="openstack-operators/4d9ed08db2179c6d86e46659534222f77f18c6c6d13f1f155004bdcf89l5n8z" Dec 06 03:52:28 crc kubenswrapper[4980]: I1206 03:52:28.121787 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/6674a828-934e-4b14-98be-740dad507b95-util\") pod \"4d9ed08db2179c6d86e46659534222f77f18c6c6d13f1f155004bdcf89l5n8z\" (UID: \"6674a828-934e-4b14-98be-740dad507b95\") " pod="openstack-operators/4d9ed08db2179c6d86e46659534222f77f18c6c6d13f1f155004bdcf89l5n8z" Dec 06 03:52:28 crc kubenswrapper[4980]: I1206 03:52:28.222989 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4hx4m\" (UniqueName: \"kubernetes.io/projected/6674a828-934e-4b14-98be-740dad507b95-kube-api-access-4hx4m\") pod \"4d9ed08db2179c6d86e46659534222f77f18c6c6d13f1f155004bdcf89l5n8z\" (UID: \"6674a828-934e-4b14-98be-740dad507b95\") " pod="openstack-operators/4d9ed08db2179c6d86e46659534222f77f18c6c6d13f1f155004bdcf89l5n8z" Dec 06 03:52:28 crc kubenswrapper[4980]: I1206 03:52:28.223070 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/6674a828-934e-4b14-98be-740dad507b95-bundle\") pod \"4d9ed08db2179c6d86e46659534222f77f18c6c6d13f1f155004bdcf89l5n8z\" (UID: \"6674a828-934e-4b14-98be-740dad507b95\") " pod="openstack-operators/4d9ed08db2179c6d86e46659534222f77f18c6c6d13f1f155004bdcf89l5n8z" Dec 06 03:52:28 crc kubenswrapper[4980]: I1206 03:52:28.223133 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/6674a828-934e-4b14-98be-740dad507b95-util\") pod \"4d9ed08db2179c6d86e46659534222f77f18c6c6d13f1f155004bdcf89l5n8z\" (UID: \"6674a828-934e-4b14-98be-740dad507b95\") " pod="openstack-operators/4d9ed08db2179c6d86e46659534222f77f18c6c6d13f1f155004bdcf89l5n8z" Dec 06 03:52:28 crc kubenswrapper[4980]: I1206 03:52:28.223503 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/6674a828-934e-4b14-98be-740dad507b95-bundle\") pod \"4d9ed08db2179c6d86e46659534222f77f18c6c6d13f1f155004bdcf89l5n8z\" (UID: \"6674a828-934e-4b14-98be-740dad507b95\") " pod="openstack-operators/4d9ed08db2179c6d86e46659534222f77f18c6c6d13f1f155004bdcf89l5n8z" Dec 06 03:52:28 crc kubenswrapper[4980]: I1206 03:52:28.223971 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/6674a828-934e-4b14-98be-740dad507b95-util\") pod \"4d9ed08db2179c6d86e46659534222f77f18c6c6d13f1f155004bdcf89l5n8z\" (UID: \"6674a828-934e-4b14-98be-740dad507b95\") " pod="openstack-operators/4d9ed08db2179c6d86e46659534222f77f18c6c6d13f1f155004bdcf89l5n8z" Dec 06 03:52:28 crc kubenswrapper[4980]: I1206 03:52:28.241375 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4hx4m\" (UniqueName: \"kubernetes.io/projected/6674a828-934e-4b14-98be-740dad507b95-kube-api-access-4hx4m\") pod \"4d9ed08db2179c6d86e46659534222f77f18c6c6d13f1f155004bdcf89l5n8z\" (UID: \"6674a828-934e-4b14-98be-740dad507b95\") " pod="openstack-operators/4d9ed08db2179c6d86e46659534222f77f18c6c6d13f1f155004bdcf89l5n8z" Dec 06 03:52:28 crc kubenswrapper[4980]: I1206 03:52:28.441569 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/4d9ed08db2179c6d86e46659534222f77f18c6c6d13f1f155004bdcf89l5n8z" Dec 06 03:52:28 crc kubenswrapper[4980]: I1206 03:52:28.585956 4980 generic.go:334] "Generic (PLEG): container finished" podID="b59d13fa-5f2a-4f67-9502-4b81c035ee10" containerID="0c9b64e989e4b5c8bd068fc3fa68a9dddb214873631d9522df5db71002e96622" exitCode=0 Dec 06 03:52:28 crc kubenswrapper[4980]: I1206 03:52:28.586181 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/c03115aeb6cae4480b1000a58090a522fd8798c774d18bf66e63c4d3d15vssx" event={"ID":"b59d13fa-5f2a-4f67-9502-4b81c035ee10","Type":"ContainerDied","Data":"0c9b64e989e4b5c8bd068fc3fa68a9dddb214873631d9522df5db71002e96622"} Dec 06 03:52:28 crc kubenswrapper[4980]: I1206 03:52:28.586352 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/c03115aeb6cae4480b1000a58090a522fd8798c774d18bf66e63c4d3d15vssx" event={"ID":"b59d13fa-5f2a-4f67-9502-4b81c035ee10","Type":"ContainerStarted","Data":"9accdff1bea9bb735a228663d4e9d04320147f5d259160eba9b1721d9ac3ee02"} Dec 06 03:52:28 crc kubenswrapper[4980]: I1206 03:52:28.589287 4980 generic.go:334] "Generic (PLEG): container finished" podID="d8d8646d-f06f-403c-9ed7-e870564c7329" containerID="3e4480c29db3d34e5961de14acf7a1876fe97eb74b5a64c421b4e09e932d21ae" exitCode=0 Dec 06 03:52:28 crc kubenswrapper[4980]: I1206 03:52:28.589329 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/keystone-db-sync-5bvrm" event={"ID":"d8d8646d-f06f-403c-9ed7-e870564c7329","Type":"ContainerDied","Data":"3e4480c29db3d34e5961de14acf7a1876fe97eb74b5a64c421b4e09e932d21ae"} Dec 06 03:52:28 crc kubenswrapper[4980]: I1206 03:52:28.896656 4980 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/4d9ed08db2179c6d86e46659534222f77f18c6c6d13f1f155004bdcf89l5n8z"] Dec 06 03:52:29 crc kubenswrapper[4980]: I1206 03:52:29.599650 4980 generic.go:334] "Generic (PLEG): container finished" podID="6674a828-934e-4b14-98be-740dad507b95" containerID="ddc2c115f7609d7dbb89981cac759beed087eb45707b221aa51e859ab7a5cf7d" exitCode=0 Dec 06 03:52:29 crc kubenswrapper[4980]: I1206 03:52:29.599741 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/4d9ed08db2179c6d86e46659534222f77f18c6c6d13f1f155004bdcf89l5n8z" event={"ID":"6674a828-934e-4b14-98be-740dad507b95","Type":"ContainerDied","Data":"ddc2c115f7609d7dbb89981cac759beed087eb45707b221aa51e859ab7a5cf7d"} Dec 06 03:52:29 crc kubenswrapper[4980]: I1206 03:52:29.600047 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/4d9ed08db2179c6d86e46659534222f77f18c6c6d13f1f155004bdcf89l5n8z" event={"ID":"6674a828-934e-4b14-98be-740dad507b95","Type":"ContainerStarted","Data":"8e45bda66bb1274d927168786d61aec30b903fa8feb0521aff7814c3060b408f"} Dec 06 03:52:29 crc kubenswrapper[4980]: I1206 03:52:29.647160 4980 generic.go:334] "Generic (PLEG): container finished" podID="b59d13fa-5f2a-4f67-9502-4b81c035ee10" containerID="016af686c101b99ffb545e6ece2e91d0d9355aed520b64ba265552805c836dad" exitCode=0 Dec 06 03:52:29 crc kubenswrapper[4980]: I1206 03:52:29.647234 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/c03115aeb6cae4480b1000a58090a522fd8798c774d18bf66e63c4d3d15vssx" event={"ID":"b59d13fa-5f2a-4f67-9502-4b81c035ee10","Type":"ContainerDied","Data":"016af686c101b99ffb545e6ece2e91d0d9355aed520b64ba265552805c836dad"} Dec 06 03:52:29 crc kubenswrapper[4980]: I1206 03:52:29.955636 4980 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/keystone-db-sync-5bvrm" Dec 06 03:52:30 crc kubenswrapper[4980]: I1206 03:52:30.151234 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d8d8646d-f06f-403c-9ed7-e870564c7329-config-data\") pod \"d8d8646d-f06f-403c-9ed7-e870564c7329\" (UID: \"d8d8646d-f06f-403c-9ed7-e870564c7329\") " Dec 06 03:52:30 crc kubenswrapper[4980]: I1206 03:52:30.151800 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fmpvp\" (UniqueName: \"kubernetes.io/projected/d8d8646d-f06f-403c-9ed7-e870564c7329-kube-api-access-fmpvp\") pod \"d8d8646d-f06f-403c-9ed7-e870564c7329\" (UID: \"d8d8646d-f06f-403c-9ed7-e870564c7329\") " Dec 06 03:52:30 crc kubenswrapper[4980]: I1206 03:52:30.157347 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d8d8646d-f06f-403c-9ed7-e870564c7329-kube-api-access-fmpvp" (OuterVolumeSpecName: "kube-api-access-fmpvp") pod "d8d8646d-f06f-403c-9ed7-e870564c7329" (UID: "d8d8646d-f06f-403c-9ed7-e870564c7329"). InnerVolumeSpecName "kube-api-access-fmpvp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 03:52:30 crc kubenswrapper[4980]: I1206 03:52:30.209706 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d8d8646d-f06f-403c-9ed7-e870564c7329-config-data" (OuterVolumeSpecName: "config-data") pod "d8d8646d-f06f-403c-9ed7-e870564c7329" (UID: "d8d8646d-f06f-403c-9ed7-e870564c7329"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 03:52:30 crc kubenswrapper[4980]: I1206 03:52:30.253794 4980 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d8d8646d-f06f-403c-9ed7-e870564c7329-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 03:52:30 crc kubenswrapper[4980]: I1206 03:52:30.253850 4980 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fmpvp\" (UniqueName: \"kubernetes.io/projected/d8d8646d-f06f-403c-9ed7-e870564c7329-kube-api-access-fmpvp\") on node \"crc\" DevicePath \"\"" Dec 06 03:52:30 crc kubenswrapper[4980]: I1206 03:52:30.654816 4980 generic.go:334] "Generic (PLEG): container finished" podID="6674a828-934e-4b14-98be-740dad507b95" containerID="520543b1b0a234c03ddc5374011a1cc1cf5ebd23bf51c950d493fd41679e6e72" exitCode=0 Dec 06 03:52:30 crc kubenswrapper[4980]: I1206 03:52:30.654902 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/4d9ed08db2179c6d86e46659534222f77f18c6c6d13f1f155004bdcf89l5n8z" event={"ID":"6674a828-934e-4b14-98be-740dad507b95","Type":"ContainerDied","Data":"520543b1b0a234c03ddc5374011a1cc1cf5ebd23bf51c950d493fd41679e6e72"} Dec 06 03:52:30 crc kubenswrapper[4980]: I1206 03:52:30.659978 4980 generic.go:334] "Generic (PLEG): container finished" podID="b59d13fa-5f2a-4f67-9502-4b81c035ee10" containerID="f33e74afec7ca5319342cfac244937746d9909d3d10b95621de0481e13ac2b04" exitCode=0 Dec 06 03:52:30 crc kubenswrapper[4980]: I1206 03:52:30.660043 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/c03115aeb6cae4480b1000a58090a522fd8798c774d18bf66e63c4d3d15vssx" event={"ID":"b59d13fa-5f2a-4f67-9502-4b81c035ee10","Type":"ContainerDied","Data":"f33e74afec7ca5319342cfac244937746d9909d3d10b95621de0481e13ac2b04"} Dec 06 03:52:30 crc kubenswrapper[4980]: I1206 03:52:30.661868 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/keystone-db-sync-5bvrm" event={"ID":"d8d8646d-f06f-403c-9ed7-e870564c7329","Type":"ContainerDied","Data":"aa77285cd65a07f72c6cd39d205e9db53c986d8a6344e85cc643379efedd8cb6"} Dec 06 03:52:30 crc kubenswrapper[4980]: I1206 03:52:30.661909 4980 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="aa77285cd65a07f72c6cd39d205e9db53c986d8a6344e85cc643379efedd8cb6" Dec 06 03:52:30 crc kubenswrapper[4980]: I1206 03:52:30.661942 4980 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/keystone-db-sync-5bvrm" Dec 06 03:52:30 crc kubenswrapper[4980]: I1206 03:52:30.851704 4980 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/keystone-bootstrap-4lh5q"] Dec 06 03:52:30 crc kubenswrapper[4980]: E1206 03:52:30.852276 4980 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d8d8646d-f06f-403c-9ed7-e870564c7329" containerName="keystone-db-sync" Dec 06 03:52:30 crc kubenswrapper[4980]: I1206 03:52:30.852297 4980 state_mem.go:107] "Deleted CPUSet assignment" podUID="d8d8646d-f06f-403c-9ed7-e870564c7329" containerName="keystone-db-sync" Dec 06 03:52:30 crc kubenswrapper[4980]: I1206 03:52:30.852477 4980 memory_manager.go:354] "RemoveStaleState removing state" podUID="d8d8646d-f06f-403c-9ed7-e870564c7329" containerName="keystone-db-sync" Dec 06 03:52:30 crc kubenswrapper[4980]: I1206 03:52:30.853043 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/keystone-bootstrap-4lh5q" Dec 06 03:52:30 crc kubenswrapper[4980]: I1206 03:52:30.855594 4980 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"keystone" Dec 06 03:52:30 crc kubenswrapper[4980]: I1206 03:52:30.855653 4980 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"keystone-keystone-dockercfg-hsvtv" Dec 06 03:52:30 crc kubenswrapper[4980]: I1206 03:52:30.855778 4980 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"keystone-scripts" Dec 06 03:52:30 crc kubenswrapper[4980]: I1206 03:52:30.855937 4980 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"keystone-config-data" Dec 06 03:52:30 crc kubenswrapper[4980]: I1206 03:52:30.862410 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/01a2d5bc-6b08-4407-b9c4-e588a789cccf-credential-keys\") pod \"keystone-bootstrap-4lh5q\" (UID: \"01a2d5bc-6b08-4407-b9c4-e588a789cccf\") " pod="glance-kuttl-tests/keystone-bootstrap-4lh5q" Dec 06 03:52:30 crc kubenswrapper[4980]: I1206 03:52:30.862488 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/01a2d5bc-6b08-4407-b9c4-e588a789cccf-scripts\") pod \"keystone-bootstrap-4lh5q\" (UID: \"01a2d5bc-6b08-4407-b9c4-e588a789cccf\") " pod="glance-kuttl-tests/keystone-bootstrap-4lh5q" Dec 06 03:52:30 crc kubenswrapper[4980]: I1206 03:52:30.862596 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/01a2d5bc-6b08-4407-b9c4-e588a789cccf-config-data\") pod \"keystone-bootstrap-4lh5q\" (UID: \"01a2d5bc-6b08-4407-b9c4-e588a789cccf\") " pod="glance-kuttl-tests/keystone-bootstrap-4lh5q" Dec 06 03:52:30 crc kubenswrapper[4980]: I1206 03:52:30.862648 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n5489\" (UniqueName: \"kubernetes.io/projected/01a2d5bc-6b08-4407-b9c4-e588a789cccf-kube-api-access-n5489\") pod \"keystone-bootstrap-4lh5q\" (UID: \"01a2d5bc-6b08-4407-b9c4-e588a789cccf\") " pod="glance-kuttl-tests/keystone-bootstrap-4lh5q" Dec 06 03:52:30 crc kubenswrapper[4980]: I1206 03:52:30.862688 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/01a2d5bc-6b08-4407-b9c4-e588a789cccf-fernet-keys\") pod \"keystone-bootstrap-4lh5q\" (UID: \"01a2d5bc-6b08-4407-b9c4-e588a789cccf\") " pod="glance-kuttl-tests/keystone-bootstrap-4lh5q" Dec 06 03:52:30 crc kubenswrapper[4980]: I1206 03:52:30.870845 4980 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/keystone-bootstrap-4lh5q"] Dec 06 03:52:30 crc kubenswrapper[4980]: I1206 03:52:30.963333 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/01a2d5bc-6b08-4407-b9c4-e588a789cccf-credential-keys\") pod \"keystone-bootstrap-4lh5q\" (UID: \"01a2d5bc-6b08-4407-b9c4-e588a789cccf\") " pod="glance-kuttl-tests/keystone-bootstrap-4lh5q" Dec 06 03:52:30 crc kubenswrapper[4980]: I1206 03:52:30.963499 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/01a2d5bc-6b08-4407-b9c4-e588a789cccf-scripts\") pod \"keystone-bootstrap-4lh5q\" (UID: \"01a2d5bc-6b08-4407-b9c4-e588a789cccf\") " pod="glance-kuttl-tests/keystone-bootstrap-4lh5q" Dec 06 03:52:30 crc kubenswrapper[4980]: I1206 03:52:30.963675 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/01a2d5bc-6b08-4407-b9c4-e588a789cccf-config-data\") pod \"keystone-bootstrap-4lh5q\" (UID: \"01a2d5bc-6b08-4407-b9c4-e588a789cccf\") " pod="glance-kuttl-tests/keystone-bootstrap-4lh5q" Dec 06 03:52:30 crc kubenswrapper[4980]: I1206 03:52:30.963798 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n5489\" (UniqueName: \"kubernetes.io/projected/01a2d5bc-6b08-4407-b9c4-e588a789cccf-kube-api-access-n5489\") pod \"keystone-bootstrap-4lh5q\" (UID: \"01a2d5bc-6b08-4407-b9c4-e588a789cccf\") " pod="glance-kuttl-tests/keystone-bootstrap-4lh5q" Dec 06 03:52:30 crc kubenswrapper[4980]: I1206 03:52:30.963930 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/01a2d5bc-6b08-4407-b9c4-e588a789cccf-fernet-keys\") pod \"keystone-bootstrap-4lh5q\" (UID: \"01a2d5bc-6b08-4407-b9c4-e588a789cccf\") " pod="glance-kuttl-tests/keystone-bootstrap-4lh5q" Dec 06 03:52:30 crc kubenswrapper[4980]: I1206 03:52:30.967929 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/01a2d5bc-6b08-4407-b9c4-e588a789cccf-scripts\") pod \"keystone-bootstrap-4lh5q\" (UID: \"01a2d5bc-6b08-4407-b9c4-e588a789cccf\") " pod="glance-kuttl-tests/keystone-bootstrap-4lh5q" Dec 06 03:52:30 crc kubenswrapper[4980]: I1206 03:52:30.969021 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/01a2d5bc-6b08-4407-b9c4-e588a789cccf-config-data\") pod \"keystone-bootstrap-4lh5q\" (UID: \"01a2d5bc-6b08-4407-b9c4-e588a789cccf\") " pod="glance-kuttl-tests/keystone-bootstrap-4lh5q" Dec 06 03:52:30 crc kubenswrapper[4980]: I1206 03:52:30.969849 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/01a2d5bc-6b08-4407-b9c4-e588a789cccf-fernet-keys\") pod \"keystone-bootstrap-4lh5q\" (UID: \"01a2d5bc-6b08-4407-b9c4-e588a789cccf\") " pod="glance-kuttl-tests/keystone-bootstrap-4lh5q" Dec 06 03:52:30 crc kubenswrapper[4980]: I1206 03:52:30.970566 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/01a2d5bc-6b08-4407-b9c4-e588a789cccf-credential-keys\") pod \"keystone-bootstrap-4lh5q\" (UID: \"01a2d5bc-6b08-4407-b9c4-e588a789cccf\") " pod="glance-kuttl-tests/keystone-bootstrap-4lh5q" Dec 06 03:52:30 crc kubenswrapper[4980]: I1206 03:52:30.999359 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n5489\" (UniqueName: \"kubernetes.io/projected/01a2d5bc-6b08-4407-b9c4-e588a789cccf-kube-api-access-n5489\") pod \"keystone-bootstrap-4lh5q\" (UID: \"01a2d5bc-6b08-4407-b9c4-e588a789cccf\") " pod="glance-kuttl-tests/keystone-bootstrap-4lh5q" Dec 06 03:52:31 crc kubenswrapper[4980]: I1206 03:52:31.168101 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/keystone-bootstrap-4lh5q" Dec 06 03:52:31 crc kubenswrapper[4980]: I1206 03:52:31.627623 4980 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/keystone-bootstrap-4lh5q"] Dec 06 03:52:31 crc kubenswrapper[4980]: I1206 03:52:31.673342 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/keystone-bootstrap-4lh5q" event={"ID":"01a2d5bc-6b08-4407-b9c4-e588a789cccf","Type":"ContainerStarted","Data":"c8deebcb3ba29566b6ab3f4b9bad1ac022610ac30c80ef575c39d10f90e04871"} Dec 06 03:52:31 crc kubenswrapper[4980]: I1206 03:52:31.695314 4980 generic.go:334] "Generic (PLEG): container finished" podID="6674a828-934e-4b14-98be-740dad507b95" containerID="46827ec2a2c3c02dcfc9fc751f1dd844bb68073fb642d561f7bd4911dce5ec4f" exitCode=0 Dec 06 03:52:31 crc kubenswrapper[4980]: I1206 03:52:31.695443 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/4d9ed08db2179c6d86e46659534222f77f18c6c6d13f1f155004bdcf89l5n8z" event={"ID":"6674a828-934e-4b14-98be-740dad507b95","Type":"ContainerDied","Data":"46827ec2a2c3c02dcfc9fc751f1dd844bb68073fb642d561f7bd4911dce5ec4f"} Dec 06 03:52:31 crc kubenswrapper[4980]: I1206 03:52:31.948801 4980 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/c03115aeb6cae4480b1000a58090a522fd8798c774d18bf66e63c4d3d15vssx" Dec 06 03:52:32 crc kubenswrapper[4980]: I1206 03:52:32.080595 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/b59d13fa-5f2a-4f67-9502-4b81c035ee10-bundle\") pod \"b59d13fa-5f2a-4f67-9502-4b81c035ee10\" (UID: \"b59d13fa-5f2a-4f67-9502-4b81c035ee10\") " Dec 06 03:52:32 crc kubenswrapper[4980]: I1206 03:52:32.080684 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mp6d8\" (UniqueName: \"kubernetes.io/projected/b59d13fa-5f2a-4f67-9502-4b81c035ee10-kube-api-access-mp6d8\") pod \"b59d13fa-5f2a-4f67-9502-4b81c035ee10\" (UID: \"b59d13fa-5f2a-4f67-9502-4b81c035ee10\") " Dec 06 03:52:32 crc kubenswrapper[4980]: I1206 03:52:32.080743 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/b59d13fa-5f2a-4f67-9502-4b81c035ee10-util\") pod \"b59d13fa-5f2a-4f67-9502-4b81c035ee10\" (UID: \"b59d13fa-5f2a-4f67-9502-4b81c035ee10\") " Dec 06 03:52:32 crc kubenswrapper[4980]: I1206 03:52:32.081551 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b59d13fa-5f2a-4f67-9502-4b81c035ee10-bundle" (OuterVolumeSpecName: "bundle") pod "b59d13fa-5f2a-4f67-9502-4b81c035ee10" (UID: "b59d13fa-5f2a-4f67-9502-4b81c035ee10"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 03:52:32 crc kubenswrapper[4980]: I1206 03:52:32.092768 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b59d13fa-5f2a-4f67-9502-4b81c035ee10-kube-api-access-mp6d8" (OuterVolumeSpecName: "kube-api-access-mp6d8") pod "b59d13fa-5f2a-4f67-9502-4b81c035ee10" (UID: "b59d13fa-5f2a-4f67-9502-4b81c035ee10"). InnerVolumeSpecName "kube-api-access-mp6d8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 03:52:32 crc kubenswrapper[4980]: I1206 03:52:32.094747 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b59d13fa-5f2a-4f67-9502-4b81c035ee10-util" (OuterVolumeSpecName: "util") pod "b59d13fa-5f2a-4f67-9502-4b81c035ee10" (UID: "b59d13fa-5f2a-4f67-9502-4b81c035ee10"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 03:52:32 crc kubenswrapper[4980]: I1206 03:52:32.182568 4980 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/b59d13fa-5f2a-4f67-9502-4b81c035ee10-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 03:52:32 crc kubenswrapper[4980]: I1206 03:52:32.182611 4980 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mp6d8\" (UniqueName: \"kubernetes.io/projected/b59d13fa-5f2a-4f67-9502-4b81c035ee10-kube-api-access-mp6d8\") on node \"crc\" DevicePath \"\"" Dec 06 03:52:32 crc kubenswrapper[4980]: I1206 03:52:32.182623 4980 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/b59d13fa-5f2a-4f67-9502-4b81c035ee10-util\") on node \"crc\" DevicePath \"\"" Dec 06 03:52:32 crc kubenswrapper[4980]: I1206 03:52:32.704118 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/keystone-bootstrap-4lh5q" event={"ID":"01a2d5bc-6b08-4407-b9c4-e588a789cccf","Type":"ContainerStarted","Data":"7fec5567a90eb550b62de633c0e2046fb64e4867f89abde77306a213ceb0370f"} Dec 06 03:52:32 crc kubenswrapper[4980]: I1206 03:52:32.708441 4980 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/c03115aeb6cae4480b1000a58090a522fd8798c774d18bf66e63c4d3d15vssx" Dec 06 03:52:32 crc kubenswrapper[4980]: I1206 03:52:32.709362 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/c03115aeb6cae4480b1000a58090a522fd8798c774d18bf66e63c4d3d15vssx" event={"ID":"b59d13fa-5f2a-4f67-9502-4b81c035ee10","Type":"ContainerDied","Data":"9accdff1bea9bb735a228663d4e9d04320147f5d259160eba9b1721d9ac3ee02"} Dec 06 03:52:32 crc kubenswrapper[4980]: I1206 03:52:32.709441 4980 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9accdff1bea9bb735a228663d4e9d04320147f5d259160eba9b1721d9ac3ee02" Dec 06 03:52:32 crc kubenswrapper[4980]: I1206 03:52:32.724546 4980 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/keystone-bootstrap-4lh5q" podStartSLOduration=2.724505432 podStartE2EDuration="2.724505432s" podCreationTimestamp="2025-12-06 03:52:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 03:52:32.722784472 +0000 UTC m=+1111.967535753" watchObservedRunningTime="2025-12-06 03:52:32.724505432 +0000 UTC m=+1111.969256703" Dec 06 03:52:33 crc kubenswrapper[4980]: I1206 03:52:32.999124 4980 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/4d9ed08db2179c6d86e46659534222f77f18c6c6d13f1f155004bdcf89l5n8z" Dec 06 03:52:33 crc kubenswrapper[4980]: I1206 03:52:33.096039 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/6674a828-934e-4b14-98be-740dad507b95-util\") pod \"6674a828-934e-4b14-98be-740dad507b95\" (UID: \"6674a828-934e-4b14-98be-740dad507b95\") " Dec 06 03:52:33 crc kubenswrapper[4980]: I1206 03:52:33.096417 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/6674a828-934e-4b14-98be-740dad507b95-bundle\") pod \"6674a828-934e-4b14-98be-740dad507b95\" (UID: \"6674a828-934e-4b14-98be-740dad507b95\") " Dec 06 03:52:33 crc kubenswrapper[4980]: I1206 03:52:33.096449 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4hx4m\" (UniqueName: \"kubernetes.io/projected/6674a828-934e-4b14-98be-740dad507b95-kube-api-access-4hx4m\") pod \"6674a828-934e-4b14-98be-740dad507b95\" (UID: \"6674a828-934e-4b14-98be-740dad507b95\") " Dec 06 03:52:33 crc kubenswrapper[4980]: I1206 03:52:33.098133 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6674a828-934e-4b14-98be-740dad507b95-bundle" (OuterVolumeSpecName: "bundle") pod "6674a828-934e-4b14-98be-740dad507b95" (UID: "6674a828-934e-4b14-98be-740dad507b95"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 03:52:33 crc kubenswrapper[4980]: I1206 03:52:33.114000 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6674a828-934e-4b14-98be-740dad507b95-util" (OuterVolumeSpecName: "util") pod "6674a828-934e-4b14-98be-740dad507b95" (UID: "6674a828-934e-4b14-98be-740dad507b95"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 03:52:33 crc kubenswrapper[4980]: I1206 03:52:33.118334 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6674a828-934e-4b14-98be-740dad507b95-kube-api-access-4hx4m" (OuterVolumeSpecName: "kube-api-access-4hx4m") pod "6674a828-934e-4b14-98be-740dad507b95" (UID: "6674a828-934e-4b14-98be-740dad507b95"). InnerVolumeSpecName "kube-api-access-4hx4m". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 03:52:33 crc kubenswrapper[4980]: I1206 03:52:33.200359 4980 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/6674a828-934e-4b14-98be-740dad507b95-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 03:52:33 crc kubenswrapper[4980]: I1206 03:52:33.200401 4980 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4hx4m\" (UniqueName: \"kubernetes.io/projected/6674a828-934e-4b14-98be-740dad507b95-kube-api-access-4hx4m\") on node \"crc\" DevicePath \"\"" Dec 06 03:52:33 crc kubenswrapper[4980]: I1206 03:52:33.200422 4980 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/6674a828-934e-4b14-98be-740dad507b95-util\") on node \"crc\" DevicePath \"\"" Dec 06 03:52:33 crc kubenswrapper[4980]: I1206 03:52:33.716396 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/4d9ed08db2179c6d86e46659534222f77f18c6c6d13f1f155004bdcf89l5n8z" event={"ID":"6674a828-934e-4b14-98be-740dad507b95","Type":"ContainerDied","Data":"8e45bda66bb1274d927168786d61aec30b903fa8feb0521aff7814c3060b408f"} Dec 06 03:52:33 crc kubenswrapper[4980]: I1206 03:52:33.716440 4980 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8e45bda66bb1274d927168786d61aec30b903fa8feb0521aff7814c3060b408f" Dec 06 03:52:33 crc kubenswrapper[4980]: I1206 03:52:33.716491 4980 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/4d9ed08db2179c6d86e46659534222f77f18c6c6d13f1f155004bdcf89l5n8z" Dec 06 03:52:35 crc kubenswrapper[4980]: I1206 03:52:35.734021 4980 generic.go:334] "Generic (PLEG): container finished" podID="01a2d5bc-6b08-4407-b9c4-e588a789cccf" containerID="7fec5567a90eb550b62de633c0e2046fb64e4867f89abde77306a213ceb0370f" exitCode=0 Dec 06 03:52:35 crc kubenswrapper[4980]: I1206 03:52:35.734118 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/keystone-bootstrap-4lh5q" event={"ID":"01a2d5bc-6b08-4407-b9c4-e588a789cccf","Type":"ContainerDied","Data":"7fec5567a90eb550b62de633c0e2046fb64e4867f89abde77306a213ceb0370f"} Dec 06 03:52:37 crc kubenswrapper[4980]: I1206 03:52:37.055447 4980 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/keystone-bootstrap-4lh5q" Dec 06 03:52:37 crc kubenswrapper[4980]: I1206 03:52:37.158159 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/01a2d5bc-6b08-4407-b9c4-e588a789cccf-credential-keys\") pod \"01a2d5bc-6b08-4407-b9c4-e588a789cccf\" (UID: \"01a2d5bc-6b08-4407-b9c4-e588a789cccf\") " Dec 06 03:52:37 crc kubenswrapper[4980]: I1206 03:52:37.158304 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n5489\" (UniqueName: \"kubernetes.io/projected/01a2d5bc-6b08-4407-b9c4-e588a789cccf-kube-api-access-n5489\") pod \"01a2d5bc-6b08-4407-b9c4-e588a789cccf\" (UID: \"01a2d5bc-6b08-4407-b9c4-e588a789cccf\") " Dec 06 03:52:37 crc kubenswrapper[4980]: I1206 03:52:37.158327 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/01a2d5bc-6b08-4407-b9c4-e588a789cccf-fernet-keys\") pod \"01a2d5bc-6b08-4407-b9c4-e588a789cccf\" (UID: \"01a2d5bc-6b08-4407-b9c4-e588a789cccf\") " Dec 06 03:52:37 crc kubenswrapper[4980]: I1206 03:52:37.161273 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/01a2d5bc-6b08-4407-b9c4-e588a789cccf-scripts\") pod \"01a2d5bc-6b08-4407-b9c4-e588a789cccf\" (UID: \"01a2d5bc-6b08-4407-b9c4-e588a789cccf\") " Dec 06 03:52:37 crc kubenswrapper[4980]: I1206 03:52:37.161313 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/01a2d5bc-6b08-4407-b9c4-e588a789cccf-config-data\") pod \"01a2d5bc-6b08-4407-b9c4-e588a789cccf\" (UID: \"01a2d5bc-6b08-4407-b9c4-e588a789cccf\") " Dec 06 03:52:37 crc kubenswrapper[4980]: I1206 03:52:37.170659 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01a2d5bc-6b08-4407-b9c4-e588a789cccf-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "01a2d5bc-6b08-4407-b9c4-e588a789cccf" (UID: "01a2d5bc-6b08-4407-b9c4-e588a789cccf"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 03:52:37 crc kubenswrapper[4980]: I1206 03:52:37.171057 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01a2d5bc-6b08-4407-b9c4-e588a789cccf-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "01a2d5bc-6b08-4407-b9c4-e588a789cccf" (UID: "01a2d5bc-6b08-4407-b9c4-e588a789cccf"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 03:52:37 crc kubenswrapper[4980]: I1206 03:52:37.171198 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/01a2d5bc-6b08-4407-b9c4-e588a789cccf-kube-api-access-n5489" (OuterVolumeSpecName: "kube-api-access-n5489") pod "01a2d5bc-6b08-4407-b9c4-e588a789cccf" (UID: "01a2d5bc-6b08-4407-b9c4-e588a789cccf"). InnerVolumeSpecName "kube-api-access-n5489". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 03:52:37 crc kubenswrapper[4980]: I1206 03:52:37.171837 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01a2d5bc-6b08-4407-b9c4-e588a789cccf-scripts" (OuterVolumeSpecName: "scripts") pod "01a2d5bc-6b08-4407-b9c4-e588a789cccf" (UID: "01a2d5bc-6b08-4407-b9c4-e588a789cccf"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 03:52:37 crc kubenswrapper[4980]: I1206 03:52:37.194780 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01a2d5bc-6b08-4407-b9c4-e588a789cccf-config-data" (OuterVolumeSpecName: "config-data") pod "01a2d5bc-6b08-4407-b9c4-e588a789cccf" (UID: "01a2d5bc-6b08-4407-b9c4-e588a789cccf"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 03:52:37 crc kubenswrapper[4980]: I1206 03:52:37.263764 4980 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/01a2d5bc-6b08-4407-b9c4-e588a789cccf-credential-keys\") on node \"crc\" DevicePath \"\"" Dec 06 03:52:37 crc kubenswrapper[4980]: I1206 03:52:37.264206 4980 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n5489\" (UniqueName: \"kubernetes.io/projected/01a2d5bc-6b08-4407-b9c4-e588a789cccf-kube-api-access-n5489\") on node \"crc\" DevicePath \"\"" Dec 06 03:52:37 crc kubenswrapper[4980]: I1206 03:52:37.264335 4980 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/01a2d5bc-6b08-4407-b9c4-e588a789cccf-fernet-keys\") on node \"crc\" DevicePath \"\"" Dec 06 03:52:37 crc kubenswrapper[4980]: I1206 03:52:37.264427 4980 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/01a2d5bc-6b08-4407-b9c4-e588a789cccf-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 03:52:37 crc kubenswrapper[4980]: I1206 03:52:37.264533 4980 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/01a2d5bc-6b08-4407-b9c4-e588a789cccf-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 03:52:37 crc kubenswrapper[4980]: I1206 03:52:37.751082 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/keystone-bootstrap-4lh5q" event={"ID":"01a2d5bc-6b08-4407-b9c4-e588a789cccf","Type":"ContainerDied","Data":"c8deebcb3ba29566b6ab3f4b9bad1ac022610ac30c80ef575c39d10f90e04871"} Dec 06 03:52:37 crc kubenswrapper[4980]: I1206 03:52:37.751128 4980 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c8deebcb3ba29566b6ab3f4b9bad1ac022610ac30c80ef575c39d10f90e04871" Dec 06 03:52:37 crc kubenswrapper[4980]: I1206 03:52:37.751230 4980 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/keystone-bootstrap-4lh5q" Dec 06 03:52:37 crc kubenswrapper[4980]: I1206 03:52:37.926022 4980 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/keystone-568cc6458b-84xg6"] Dec 06 03:52:37 crc kubenswrapper[4980]: E1206 03:52:37.926785 4980 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b59d13fa-5f2a-4f67-9502-4b81c035ee10" containerName="extract" Dec 06 03:52:37 crc kubenswrapper[4980]: I1206 03:52:37.926812 4980 state_mem.go:107] "Deleted CPUSet assignment" podUID="b59d13fa-5f2a-4f67-9502-4b81c035ee10" containerName="extract" Dec 06 03:52:37 crc kubenswrapper[4980]: E1206 03:52:37.926833 4980 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6674a828-934e-4b14-98be-740dad507b95" containerName="extract" Dec 06 03:52:37 crc kubenswrapper[4980]: I1206 03:52:37.926841 4980 state_mem.go:107] "Deleted CPUSet assignment" podUID="6674a828-934e-4b14-98be-740dad507b95" containerName="extract" Dec 06 03:52:37 crc kubenswrapper[4980]: E1206 03:52:37.926859 4980 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6674a828-934e-4b14-98be-740dad507b95" containerName="util" Dec 06 03:52:37 crc kubenswrapper[4980]: I1206 03:52:37.926867 4980 state_mem.go:107] "Deleted CPUSet assignment" podUID="6674a828-934e-4b14-98be-740dad507b95" containerName="util" Dec 06 03:52:37 crc kubenswrapper[4980]: E1206 03:52:37.926874 4980 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6674a828-934e-4b14-98be-740dad507b95" containerName="pull" Dec 06 03:52:37 crc kubenswrapper[4980]: I1206 03:52:37.926881 4980 state_mem.go:107] "Deleted CPUSet assignment" podUID="6674a828-934e-4b14-98be-740dad507b95" containerName="pull" Dec 06 03:52:37 crc kubenswrapper[4980]: E1206 03:52:37.926903 4980 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b59d13fa-5f2a-4f67-9502-4b81c035ee10" containerName="util" Dec 06 03:52:37 crc kubenswrapper[4980]: I1206 03:52:37.926910 4980 state_mem.go:107] "Deleted CPUSet assignment" podUID="b59d13fa-5f2a-4f67-9502-4b81c035ee10" containerName="util" Dec 06 03:52:37 crc kubenswrapper[4980]: E1206 03:52:37.926924 4980 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b59d13fa-5f2a-4f67-9502-4b81c035ee10" containerName="pull" Dec 06 03:52:37 crc kubenswrapper[4980]: I1206 03:52:37.926931 4980 state_mem.go:107] "Deleted CPUSet assignment" podUID="b59d13fa-5f2a-4f67-9502-4b81c035ee10" containerName="pull" Dec 06 03:52:37 crc kubenswrapper[4980]: E1206 03:52:37.926938 4980 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="01a2d5bc-6b08-4407-b9c4-e588a789cccf" containerName="keystone-bootstrap" Dec 06 03:52:37 crc kubenswrapper[4980]: I1206 03:52:37.926947 4980 state_mem.go:107] "Deleted CPUSet assignment" podUID="01a2d5bc-6b08-4407-b9c4-e588a789cccf" containerName="keystone-bootstrap" Dec 06 03:52:37 crc kubenswrapper[4980]: I1206 03:52:37.927084 4980 memory_manager.go:354] "RemoveStaleState removing state" podUID="b59d13fa-5f2a-4f67-9502-4b81c035ee10" containerName="extract" Dec 06 03:52:37 crc kubenswrapper[4980]: I1206 03:52:37.927104 4980 memory_manager.go:354] "RemoveStaleState removing state" podUID="01a2d5bc-6b08-4407-b9c4-e588a789cccf" containerName="keystone-bootstrap" Dec 06 03:52:37 crc kubenswrapper[4980]: I1206 03:52:37.927119 4980 memory_manager.go:354] "RemoveStaleState removing state" podUID="6674a828-934e-4b14-98be-740dad507b95" containerName="extract" Dec 06 03:52:37 crc kubenswrapper[4980]: I1206 03:52:37.927775 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/keystone-568cc6458b-84xg6" Dec 06 03:52:37 crc kubenswrapper[4980]: I1206 03:52:37.930957 4980 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"keystone-config-data" Dec 06 03:52:37 crc kubenswrapper[4980]: I1206 03:52:37.931268 4980 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"keystone-keystone-dockercfg-hsvtv" Dec 06 03:52:37 crc kubenswrapper[4980]: I1206 03:52:37.931498 4980 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"keystone" Dec 06 03:52:37 crc kubenswrapper[4980]: I1206 03:52:37.931649 4980 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"keystone-scripts" Dec 06 03:52:37 crc kubenswrapper[4980]: I1206 03:52:37.938282 4980 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/keystone-568cc6458b-84xg6"] Dec 06 03:52:37 crc kubenswrapper[4980]: I1206 03:52:37.974573 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cd336afa-48b9-4df5-9cca-12031ce13b2d-config-data\") pod \"keystone-568cc6458b-84xg6\" (UID: \"cd336afa-48b9-4df5-9cca-12031ce13b2d\") " pod="glance-kuttl-tests/keystone-568cc6458b-84xg6" Dec 06 03:52:37 crc kubenswrapper[4980]: I1206 03:52:37.974672 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/cd336afa-48b9-4df5-9cca-12031ce13b2d-credential-keys\") pod \"keystone-568cc6458b-84xg6\" (UID: \"cd336afa-48b9-4df5-9cca-12031ce13b2d\") " pod="glance-kuttl-tests/keystone-568cc6458b-84xg6" Dec 06 03:52:37 crc kubenswrapper[4980]: I1206 03:52:37.974734 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cd336afa-48b9-4df5-9cca-12031ce13b2d-scripts\") pod \"keystone-568cc6458b-84xg6\" (UID: \"cd336afa-48b9-4df5-9cca-12031ce13b2d\") " pod="glance-kuttl-tests/keystone-568cc6458b-84xg6" Dec 06 03:52:37 crc kubenswrapper[4980]: I1206 03:52:37.974778 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/cd336afa-48b9-4df5-9cca-12031ce13b2d-fernet-keys\") pod \"keystone-568cc6458b-84xg6\" (UID: \"cd336afa-48b9-4df5-9cca-12031ce13b2d\") " pod="glance-kuttl-tests/keystone-568cc6458b-84xg6" Dec 06 03:52:37 crc kubenswrapper[4980]: I1206 03:52:37.974799 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qqdtj\" (UniqueName: \"kubernetes.io/projected/cd336afa-48b9-4df5-9cca-12031ce13b2d-kube-api-access-qqdtj\") pod \"keystone-568cc6458b-84xg6\" (UID: \"cd336afa-48b9-4df5-9cca-12031ce13b2d\") " pod="glance-kuttl-tests/keystone-568cc6458b-84xg6" Dec 06 03:52:38 crc kubenswrapper[4980]: I1206 03:52:38.076070 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cd336afa-48b9-4df5-9cca-12031ce13b2d-scripts\") pod \"keystone-568cc6458b-84xg6\" (UID: \"cd336afa-48b9-4df5-9cca-12031ce13b2d\") " pod="glance-kuttl-tests/keystone-568cc6458b-84xg6" Dec 06 03:52:38 crc kubenswrapper[4980]: I1206 03:52:38.076165 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/cd336afa-48b9-4df5-9cca-12031ce13b2d-fernet-keys\") pod \"keystone-568cc6458b-84xg6\" (UID: \"cd336afa-48b9-4df5-9cca-12031ce13b2d\") " pod="glance-kuttl-tests/keystone-568cc6458b-84xg6" Dec 06 03:52:38 crc kubenswrapper[4980]: I1206 03:52:38.076202 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qqdtj\" (UniqueName: \"kubernetes.io/projected/cd336afa-48b9-4df5-9cca-12031ce13b2d-kube-api-access-qqdtj\") pod \"keystone-568cc6458b-84xg6\" (UID: \"cd336afa-48b9-4df5-9cca-12031ce13b2d\") " pod="glance-kuttl-tests/keystone-568cc6458b-84xg6" Dec 06 03:52:38 crc kubenswrapper[4980]: I1206 03:52:38.076263 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cd336afa-48b9-4df5-9cca-12031ce13b2d-config-data\") pod \"keystone-568cc6458b-84xg6\" (UID: \"cd336afa-48b9-4df5-9cca-12031ce13b2d\") " pod="glance-kuttl-tests/keystone-568cc6458b-84xg6" Dec 06 03:52:38 crc kubenswrapper[4980]: I1206 03:52:38.076355 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/cd336afa-48b9-4df5-9cca-12031ce13b2d-credential-keys\") pod \"keystone-568cc6458b-84xg6\" (UID: \"cd336afa-48b9-4df5-9cca-12031ce13b2d\") " pod="glance-kuttl-tests/keystone-568cc6458b-84xg6" Dec 06 03:52:38 crc kubenswrapper[4980]: I1206 03:52:38.080981 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cd336afa-48b9-4df5-9cca-12031ce13b2d-scripts\") pod \"keystone-568cc6458b-84xg6\" (UID: \"cd336afa-48b9-4df5-9cca-12031ce13b2d\") " pod="glance-kuttl-tests/keystone-568cc6458b-84xg6" Dec 06 03:52:38 crc kubenswrapper[4980]: I1206 03:52:38.081325 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/cd336afa-48b9-4df5-9cca-12031ce13b2d-credential-keys\") pod \"keystone-568cc6458b-84xg6\" (UID: \"cd336afa-48b9-4df5-9cca-12031ce13b2d\") " pod="glance-kuttl-tests/keystone-568cc6458b-84xg6" Dec 06 03:52:38 crc kubenswrapper[4980]: I1206 03:52:38.081523 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/cd336afa-48b9-4df5-9cca-12031ce13b2d-fernet-keys\") pod \"keystone-568cc6458b-84xg6\" (UID: \"cd336afa-48b9-4df5-9cca-12031ce13b2d\") " pod="glance-kuttl-tests/keystone-568cc6458b-84xg6" Dec 06 03:52:38 crc kubenswrapper[4980]: I1206 03:52:38.082632 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cd336afa-48b9-4df5-9cca-12031ce13b2d-config-data\") pod \"keystone-568cc6458b-84xg6\" (UID: \"cd336afa-48b9-4df5-9cca-12031ce13b2d\") " pod="glance-kuttl-tests/keystone-568cc6458b-84xg6" Dec 06 03:52:38 crc kubenswrapper[4980]: I1206 03:52:38.135593 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qqdtj\" (UniqueName: \"kubernetes.io/projected/cd336afa-48b9-4df5-9cca-12031ce13b2d-kube-api-access-qqdtj\") pod \"keystone-568cc6458b-84xg6\" (UID: \"cd336afa-48b9-4df5-9cca-12031ce13b2d\") " pod="glance-kuttl-tests/keystone-568cc6458b-84xg6" Dec 06 03:52:38 crc kubenswrapper[4980]: I1206 03:52:38.249168 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/keystone-568cc6458b-84xg6" Dec 06 03:52:38 crc kubenswrapper[4980]: I1206 03:52:38.504566 4980 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/keystone-568cc6458b-84xg6"] Dec 06 03:52:38 crc kubenswrapper[4980]: I1206 03:52:38.758647 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/keystone-568cc6458b-84xg6" event={"ID":"cd336afa-48b9-4df5-9cca-12031ce13b2d","Type":"ContainerStarted","Data":"c7f900070c0478d659d07abdf803ba82077a4eb61bfb83d2e216efadbab690ad"} Dec 06 03:52:39 crc kubenswrapper[4980]: I1206 03:52:39.769953 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/keystone-568cc6458b-84xg6" event={"ID":"cd336afa-48b9-4df5-9cca-12031ce13b2d","Type":"ContainerStarted","Data":"6faeae0f74ad9723950019d8d73b72123eeb6b061c42d2542e0a07ecca2b79a9"} Dec 06 03:52:39 crc kubenswrapper[4980]: I1206 03:52:39.771421 4980 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/keystone-568cc6458b-84xg6" Dec 06 03:52:39 crc kubenswrapper[4980]: I1206 03:52:39.791459 4980 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/keystone-568cc6458b-84xg6" podStartSLOduration=2.791416362 podStartE2EDuration="2.791416362s" podCreationTimestamp="2025-12-06 03:52:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 03:52:39.787001305 +0000 UTC m=+1119.031752616" watchObservedRunningTime="2025-12-06 03:52:39.791416362 +0000 UTC m=+1119.036167633" Dec 06 03:52:45 crc kubenswrapper[4980]: I1206 03:52:45.906854 4980 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/horizon-operator-controller-manager-558977579-kknvn"] Dec 06 03:52:45 crc kubenswrapper[4980]: I1206 03:52:45.908319 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-558977579-kknvn" Dec 06 03:52:45 crc kubenswrapper[4980]: I1206 03:52:45.910373 4980 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"horizon-operator-controller-manager-dockercfg-x28vz" Dec 06 03:52:45 crc kubenswrapper[4980]: I1206 03:52:45.910947 4980 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"horizon-operator-controller-manager-service-cert" Dec 06 03:52:45 crc kubenswrapper[4980]: I1206 03:52:45.935188 4980 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-558977579-kknvn"] Dec 06 03:52:46 crc kubenswrapper[4980]: I1206 03:52:46.085181 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/c4a8121a-cce4-481a-b758-ae1b5dd89336-webhook-cert\") pod \"horizon-operator-controller-manager-558977579-kknvn\" (UID: \"c4a8121a-cce4-481a-b758-ae1b5dd89336\") " pod="openstack-operators/horizon-operator-controller-manager-558977579-kknvn" Dec 06 03:52:46 crc kubenswrapper[4980]: I1206 03:52:46.085254 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2f5dg\" (UniqueName: \"kubernetes.io/projected/c4a8121a-cce4-481a-b758-ae1b5dd89336-kube-api-access-2f5dg\") pod \"horizon-operator-controller-manager-558977579-kknvn\" (UID: \"c4a8121a-cce4-481a-b758-ae1b5dd89336\") " pod="openstack-operators/horizon-operator-controller-manager-558977579-kknvn" Dec 06 03:52:46 crc kubenswrapper[4980]: I1206 03:52:46.085303 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/c4a8121a-cce4-481a-b758-ae1b5dd89336-apiservice-cert\") pod \"horizon-operator-controller-manager-558977579-kknvn\" (UID: \"c4a8121a-cce4-481a-b758-ae1b5dd89336\") " pod="openstack-operators/horizon-operator-controller-manager-558977579-kknvn" Dec 06 03:52:46 crc kubenswrapper[4980]: I1206 03:52:46.186188 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/c4a8121a-cce4-481a-b758-ae1b5dd89336-webhook-cert\") pod \"horizon-operator-controller-manager-558977579-kknvn\" (UID: \"c4a8121a-cce4-481a-b758-ae1b5dd89336\") " pod="openstack-operators/horizon-operator-controller-manager-558977579-kknvn" Dec 06 03:52:46 crc kubenswrapper[4980]: I1206 03:52:46.186262 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2f5dg\" (UniqueName: \"kubernetes.io/projected/c4a8121a-cce4-481a-b758-ae1b5dd89336-kube-api-access-2f5dg\") pod \"horizon-operator-controller-manager-558977579-kknvn\" (UID: \"c4a8121a-cce4-481a-b758-ae1b5dd89336\") " pod="openstack-operators/horizon-operator-controller-manager-558977579-kknvn" Dec 06 03:52:46 crc kubenswrapper[4980]: I1206 03:52:46.186304 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/c4a8121a-cce4-481a-b758-ae1b5dd89336-apiservice-cert\") pod \"horizon-operator-controller-manager-558977579-kknvn\" (UID: \"c4a8121a-cce4-481a-b758-ae1b5dd89336\") " pod="openstack-operators/horizon-operator-controller-manager-558977579-kknvn" Dec 06 03:52:46 crc kubenswrapper[4980]: I1206 03:52:46.202809 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/c4a8121a-cce4-481a-b758-ae1b5dd89336-webhook-cert\") pod \"horizon-operator-controller-manager-558977579-kknvn\" (UID: \"c4a8121a-cce4-481a-b758-ae1b5dd89336\") " pod="openstack-operators/horizon-operator-controller-manager-558977579-kknvn" Dec 06 03:52:46 crc kubenswrapper[4980]: I1206 03:52:46.203547 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/c4a8121a-cce4-481a-b758-ae1b5dd89336-apiservice-cert\") pod \"horizon-operator-controller-manager-558977579-kknvn\" (UID: \"c4a8121a-cce4-481a-b758-ae1b5dd89336\") " pod="openstack-operators/horizon-operator-controller-manager-558977579-kknvn" Dec 06 03:52:46 crc kubenswrapper[4980]: I1206 03:52:46.203718 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2f5dg\" (UniqueName: \"kubernetes.io/projected/c4a8121a-cce4-481a-b758-ae1b5dd89336-kube-api-access-2f5dg\") pod \"horizon-operator-controller-manager-558977579-kknvn\" (UID: \"c4a8121a-cce4-481a-b758-ae1b5dd89336\") " pod="openstack-operators/horizon-operator-controller-manager-558977579-kknvn" Dec 06 03:52:46 crc kubenswrapper[4980]: I1206 03:52:46.227122 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-558977579-kknvn" Dec 06 03:52:46 crc kubenswrapper[4980]: I1206 03:52:46.710130 4980 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-558977579-kknvn"] Dec 06 03:52:46 crc kubenswrapper[4980]: I1206 03:52:46.826462 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-558977579-kknvn" event={"ID":"c4a8121a-cce4-481a-b758-ae1b5dd89336","Type":"ContainerStarted","Data":"c03abff61d001b1c9c250cbc9d49ab2cad0109d8df56bb378076988b8b2e5637"} Dec 06 03:52:51 crc kubenswrapper[4980]: I1206 03:52:51.871699 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-558977579-kknvn" event={"ID":"c4a8121a-cce4-481a-b758-ae1b5dd89336","Type":"ContainerStarted","Data":"7b163e763b46d6e83a9f7ad54faf654c0e4fb296dbb96ab3f102e3eff598ac53"} Dec 06 03:52:51 crc kubenswrapper[4980]: I1206 03:52:51.872197 4980 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/horizon-operator-controller-manager-558977579-kknvn" Dec 06 03:52:51 crc kubenswrapper[4980]: I1206 03:52:51.872208 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-558977579-kknvn" event={"ID":"c4a8121a-cce4-481a-b758-ae1b5dd89336","Type":"ContainerStarted","Data":"8a8d5590d4ea2dd34426e4c825fd9548ffaf9fa11fc4bd9b49a53514de1bc211"} Dec 06 03:52:51 crc kubenswrapper[4980]: I1206 03:52:51.889797 4980 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/horizon-operator-controller-manager-558977579-kknvn" podStartSLOduration=2.455588826 podStartE2EDuration="6.889748091s" podCreationTimestamp="2025-12-06 03:52:45 +0000 UTC" firstStartedPulling="2025-12-06 03:52:46.717984799 +0000 UTC m=+1125.962736070" lastFinishedPulling="2025-12-06 03:52:51.152144064 +0000 UTC m=+1130.396895335" observedRunningTime="2025-12-06 03:52:51.88764517 +0000 UTC m=+1131.132396461" watchObservedRunningTime="2025-12-06 03:52:51.889748091 +0000 UTC m=+1131.134499362" Dec 06 03:52:53 crc kubenswrapper[4980]: I1206 03:52:53.134916 4980 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/swift-operator-controller-manager-5d8ccb4767-r2zdb"] Dec 06 03:52:53 crc kubenswrapper[4980]: I1206 03:52:53.136328 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-5d8ccb4767-r2zdb" Dec 06 03:52:53 crc kubenswrapper[4980]: I1206 03:52:53.142526 4980 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"swift-operator-controller-manager-dockercfg-svlp7" Dec 06 03:52:53 crc kubenswrapper[4980]: I1206 03:52:53.143010 4980 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"swift-operator-controller-manager-service-cert" Dec 06 03:52:53 crc kubenswrapper[4980]: I1206 03:52:53.157003 4980 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-5d8ccb4767-r2zdb"] Dec 06 03:52:53 crc kubenswrapper[4980]: I1206 03:52:53.239230 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8mvkf\" (UniqueName: \"kubernetes.io/projected/81b1b1bc-6f44-4bca-9b34-87c42f15c07e-kube-api-access-8mvkf\") pod \"swift-operator-controller-manager-5d8ccb4767-r2zdb\" (UID: \"81b1b1bc-6f44-4bca-9b34-87c42f15c07e\") " pod="openstack-operators/swift-operator-controller-manager-5d8ccb4767-r2zdb" Dec 06 03:52:53 crc kubenswrapper[4980]: I1206 03:52:53.239324 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/81b1b1bc-6f44-4bca-9b34-87c42f15c07e-apiservice-cert\") pod \"swift-operator-controller-manager-5d8ccb4767-r2zdb\" (UID: \"81b1b1bc-6f44-4bca-9b34-87c42f15c07e\") " pod="openstack-operators/swift-operator-controller-manager-5d8ccb4767-r2zdb" Dec 06 03:52:53 crc kubenswrapper[4980]: I1206 03:52:53.239351 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/81b1b1bc-6f44-4bca-9b34-87c42f15c07e-webhook-cert\") pod \"swift-operator-controller-manager-5d8ccb4767-r2zdb\" (UID: \"81b1b1bc-6f44-4bca-9b34-87c42f15c07e\") " pod="openstack-operators/swift-operator-controller-manager-5d8ccb4767-r2zdb" Dec 06 03:52:53 crc kubenswrapper[4980]: I1206 03:52:53.340733 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8mvkf\" (UniqueName: \"kubernetes.io/projected/81b1b1bc-6f44-4bca-9b34-87c42f15c07e-kube-api-access-8mvkf\") pod \"swift-operator-controller-manager-5d8ccb4767-r2zdb\" (UID: \"81b1b1bc-6f44-4bca-9b34-87c42f15c07e\") " pod="openstack-operators/swift-operator-controller-manager-5d8ccb4767-r2zdb" Dec 06 03:52:53 crc kubenswrapper[4980]: I1206 03:52:53.341175 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/81b1b1bc-6f44-4bca-9b34-87c42f15c07e-apiservice-cert\") pod \"swift-operator-controller-manager-5d8ccb4767-r2zdb\" (UID: \"81b1b1bc-6f44-4bca-9b34-87c42f15c07e\") " pod="openstack-operators/swift-operator-controller-manager-5d8ccb4767-r2zdb" Dec 06 03:52:53 crc kubenswrapper[4980]: I1206 03:52:53.341200 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/81b1b1bc-6f44-4bca-9b34-87c42f15c07e-webhook-cert\") pod \"swift-operator-controller-manager-5d8ccb4767-r2zdb\" (UID: \"81b1b1bc-6f44-4bca-9b34-87c42f15c07e\") " pod="openstack-operators/swift-operator-controller-manager-5d8ccb4767-r2zdb" Dec 06 03:52:53 crc kubenswrapper[4980]: I1206 03:52:53.350784 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/81b1b1bc-6f44-4bca-9b34-87c42f15c07e-webhook-cert\") pod \"swift-operator-controller-manager-5d8ccb4767-r2zdb\" (UID: \"81b1b1bc-6f44-4bca-9b34-87c42f15c07e\") " pod="openstack-operators/swift-operator-controller-manager-5d8ccb4767-r2zdb" Dec 06 03:52:53 crc kubenswrapper[4980]: I1206 03:52:53.351022 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/81b1b1bc-6f44-4bca-9b34-87c42f15c07e-apiservice-cert\") pod \"swift-operator-controller-manager-5d8ccb4767-r2zdb\" (UID: \"81b1b1bc-6f44-4bca-9b34-87c42f15c07e\") " pod="openstack-operators/swift-operator-controller-manager-5d8ccb4767-r2zdb" Dec 06 03:52:53 crc kubenswrapper[4980]: I1206 03:52:53.370308 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8mvkf\" (UniqueName: \"kubernetes.io/projected/81b1b1bc-6f44-4bca-9b34-87c42f15c07e-kube-api-access-8mvkf\") pod \"swift-operator-controller-manager-5d8ccb4767-r2zdb\" (UID: \"81b1b1bc-6f44-4bca-9b34-87c42f15c07e\") " pod="openstack-operators/swift-operator-controller-manager-5d8ccb4767-r2zdb" Dec 06 03:52:53 crc kubenswrapper[4980]: I1206 03:52:53.455703 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-5d8ccb4767-r2zdb" Dec 06 03:52:53 crc kubenswrapper[4980]: I1206 03:52:53.859557 4980 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-5d8ccb4767-r2zdb"] Dec 06 03:52:53 crc kubenswrapper[4980]: I1206 03:52:53.895408 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-5d8ccb4767-r2zdb" event={"ID":"81b1b1bc-6f44-4bca-9b34-87c42f15c07e","Type":"ContainerStarted","Data":"419dc044140349de05288d93a7b6041a8fffcac307506f09b4b5f13ce0cc0f3c"} Dec 06 03:52:55 crc kubenswrapper[4980]: I1206 03:52:55.137785 4980 patch_prober.go:28] interesting pod/machine-config-daemon-r5zfb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 03:52:55 crc kubenswrapper[4980]: I1206 03:52:55.137852 4980 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-r5zfb" podUID="320f44d1-a671-4a91-b328-a8b0fdd8f23a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 03:52:55 crc kubenswrapper[4980]: I1206 03:52:55.910972 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-5d8ccb4767-r2zdb" event={"ID":"81b1b1bc-6f44-4bca-9b34-87c42f15c07e","Type":"ContainerStarted","Data":"f4c9551fdbc3fabd296b2ff9ead59baa27548afcb3bd935f5c5b52492ed0df84"} Dec 06 03:52:55 crc kubenswrapper[4980]: I1206 03:52:55.911536 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-5d8ccb4767-r2zdb" event={"ID":"81b1b1bc-6f44-4bca-9b34-87c42f15c07e","Type":"ContainerStarted","Data":"43ec61d0737902197b2567390268712cbe8392d622dac1ac97ea5da3b75c6ca6"} Dec 06 03:52:55 crc kubenswrapper[4980]: I1206 03:52:55.911557 4980 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/swift-operator-controller-manager-5d8ccb4767-r2zdb" Dec 06 03:52:55 crc kubenswrapper[4980]: I1206 03:52:55.941371 4980 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/swift-operator-controller-manager-5d8ccb4767-r2zdb" podStartSLOduration=1.377678393 podStartE2EDuration="2.941349215s" podCreationTimestamp="2025-12-06 03:52:53 +0000 UTC" firstStartedPulling="2025-12-06 03:52:53.870469084 +0000 UTC m=+1133.115220355" lastFinishedPulling="2025-12-06 03:52:55.434139916 +0000 UTC m=+1134.678891177" observedRunningTime="2025-12-06 03:52:55.936837635 +0000 UTC m=+1135.181588926" watchObservedRunningTime="2025-12-06 03:52:55.941349215 +0000 UTC m=+1135.186100486" Dec 06 03:52:56 crc kubenswrapper[4980]: I1206 03:52:56.234243 4980 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/horizon-operator-controller-manager-558977579-kknvn" Dec 06 03:53:03 crc kubenswrapper[4980]: I1206 03:53:03.460400 4980 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/swift-operator-controller-manager-5d8ccb4767-r2zdb" Dec 06 03:53:09 crc kubenswrapper[4980]: I1206 03:53:09.644623 4980 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/glance-operator-index-wt2kh"] Dec 06 03:53:09 crc kubenswrapper[4980]: I1206 03:53:09.647360 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-index-wt2kh" Dec 06 03:53:09 crc kubenswrapper[4980]: I1206 03:53:09.649304 4980 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"glance-operator-index-dockercfg-g8qlt" Dec 06 03:53:09 crc kubenswrapper[4980]: I1206 03:53:09.663373 4980 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-index-wt2kh"] Dec 06 03:53:09 crc kubenswrapper[4980]: I1206 03:53:09.725828 4980 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/keystone-568cc6458b-84xg6" Dec 06 03:53:09 crc kubenswrapper[4980]: I1206 03:53:09.809858 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-txhbf\" (UniqueName: \"kubernetes.io/projected/79e40256-a186-48dc-b663-bd971d587fb9-kube-api-access-txhbf\") pod \"glance-operator-index-wt2kh\" (UID: \"79e40256-a186-48dc-b663-bd971d587fb9\") " pod="openstack-operators/glance-operator-index-wt2kh" Dec 06 03:53:09 crc kubenswrapper[4980]: I1206 03:53:09.911570 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-txhbf\" (UniqueName: \"kubernetes.io/projected/79e40256-a186-48dc-b663-bd971d587fb9-kube-api-access-txhbf\") pod \"glance-operator-index-wt2kh\" (UID: \"79e40256-a186-48dc-b663-bd971d587fb9\") " pod="openstack-operators/glance-operator-index-wt2kh" Dec 06 03:53:09 crc kubenswrapper[4980]: I1206 03:53:09.932411 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-txhbf\" (UniqueName: \"kubernetes.io/projected/79e40256-a186-48dc-b663-bd971d587fb9-kube-api-access-txhbf\") pod \"glance-operator-index-wt2kh\" (UID: \"79e40256-a186-48dc-b663-bd971d587fb9\") " pod="openstack-operators/glance-operator-index-wt2kh" Dec 06 03:53:09 crc kubenswrapper[4980]: I1206 03:53:09.964607 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-index-wt2kh" Dec 06 03:53:10 crc kubenswrapper[4980]: I1206 03:53:10.367343 4980 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-index-wt2kh"] Dec 06 03:53:10 crc kubenswrapper[4980]: W1206 03:53:10.369500 4980 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod79e40256_a186_48dc_b663_bd971d587fb9.slice/crio-2fc9c0045a48fb08bec7a6d2aa972c88f65b3127b920d96dcde44d6583df9fdf WatchSource:0}: Error finding container 2fc9c0045a48fb08bec7a6d2aa972c88f65b3127b920d96dcde44d6583df9fdf: Status 404 returned error can't find the container with id 2fc9c0045a48fb08bec7a6d2aa972c88f65b3127b920d96dcde44d6583df9fdf Dec 06 03:53:11 crc kubenswrapper[4980]: I1206 03:53:11.034542 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-index-wt2kh" event={"ID":"79e40256-a186-48dc-b663-bd971d587fb9","Type":"ContainerStarted","Data":"2fc9c0045a48fb08bec7a6d2aa972c88f65b3127b920d96dcde44d6583df9fdf"} Dec 06 03:53:14 crc kubenswrapper[4980]: I1206 03:53:14.232365 4980 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/glance-operator-index-wt2kh"] Dec 06 03:53:14 crc kubenswrapper[4980]: I1206 03:53:14.841859 4980 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/glance-operator-index-bzpqg"] Dec 06 03:53:14 crc kubenswrapper[4980]: I1206 03:53:14.842750 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-index-bzpqg" Dec 06 03:53:14 crc kubenswrapper[4980]: I1206 03:53:14.850956 4980 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-index-bzpqg"] Dec 06 03:53:14 crc kubenswrapper[4980]: I1206 03:53:14.989652 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8944c\" (UniqueName: \"kubernetes.io/projected/c88f8da4-819c-42bf-9938-60c898c6bace-kube-api-access-8944c\") pod \"glance-operator-index-bzpqg\" (UID: \"c88f8da4-819c-42bf-9938-60c898c6bace\") " pod="openstack-operators/glance-operator-index-bzpqg" Dec 06 03:53:15 crc kubenswrapper[4980]: I1206 03:53:15.091431 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8944c\" (UniqueName: \"kubernetes.io/projected/c88f8da4-819c-42bf-9938-60c898c6bace-kube-api-access-8944c\") pod \"glance-operator-index-bzpqg\" (UID: \"c88f8da4-819c-42bf-9938-60c898c6bace\") " pod="openstack-operators/glance-operator-index-bzpqg" Dec 06 03:53:15 crc kubenswrapper[4980]: I1206 03:53:15.113986 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8944c\" (UniqueName: \"kubernetes.io/projected/c88f8da4-819c-42bf-9938-60c898c6bace-kube-api-access-8944c\") pod \"glance-operator-index-bzpqg\" (UID: \"c88f8da4-819c-42bf-9938-60c898c6bace\") " pod="openstack-operators/glance-operator-index-bzpqg" Dec 06 03:53:15 crc kubenswrapper[4980]: I1206 03:53:15.160901 4980 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/swift-proxy-85fcf7dd9-sn4m2"] Dec 06 03:53:15 crc kubenswrapper[4980]: I1206 03:53:15.163200 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/swift-proxy-85fcf7dd9-sn4m2" Dec 06 03:53:15 crc kubenswrapper[4980]: I1206 03:53:15.166566 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-index-bzpqg" Dec 06 03:53:15 crc kubenswrapper[4980]: I1206 03:53:15.167538 4980 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"swift-conf" Dec 06 03:53:15 crc kubenswrapper[4980]: I1206 03:53:15.169077 4980 reflector.go:368] Caches populated for *v1.ConfigMap from object-"glance-kuttl-tests"/"swift-ring-files" Dec 06 03:53:15 crc kubenswrapper[4980]: I1206 03:53:15.169786 4980 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"swift-proxy-config-data" Dec 06 03:53:15 crc kubenswrapper[4980]: I1206 03:53:15.169793 4980 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"swift-swift-dockercfg-2x5fp" Dec 06 03:53:15 crc kubenswrapper[4980]: I1206 03:53:15.174181 4980 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/swift-proxy-85fcf7dd9-sn4m2"] Dec 06 03:53:15 crc kubenswrapper[4980]: I1206 03:53:15.294562 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/92027099-00cc-45db-bbfe-7796fe5a68c4-etc-swift\") pod \"swift-proxy-85fcf7dd9-sn4m2\" (UID: \"92027099-00cc-45db-bbfe-7796fe5a68c4\") " pod="glance-kuttl-tests/swift-proxy-85fcf7dd9-sn4m2" Dec 06 03:53:15 crc kubenswrapper[4980]: I1206 03:53:15.294799 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/92027099-00cc-45db-bbfe-7796fe5a68c4-run-httpd\") pod \"swift-proxy-85fcf7dd9-sn4m2\" (UID: \"92027099-00cc-45db-bbfe-7796fe5a68c4\") " pod="glance-kuttl-tests/swift-proxy-85fcf7dd9-sn4m2" Dec 06 03:53:15 crc kubenswrapper[4980]: I1206 03:53:15.294849 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5c9mn\" (UniqueName: \"kubernetes.io/projected/92027099-00cc-45db-bbfe-7796fe5a68c4-kube-api-access-5c9mn\") pod \"swift-proxy-85fcf7dd9-sn4m2\" (UID: \"92027099-00cc-45db-bbfe-7796fe5a68c4\") " pod="glance-kuttl-tests/swift-proxy-85fcf7dd9-sn4m2" Dec 06 03:53:15 crc kubenswrapper[4980]: I1206 03:53:15.294878 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/92027099-00cc-45db-bbfe-7796fe5a68c4-config-data\") pod \"swift-proxy-85fcf7dd9-sn4m2\" (UID: \"92027099-00cc-45db-bbfe-7796fe5a68c4\") " pod="glance-kuttl-tests/swift-proxy-85fcf7dd9-sn4m2" Dec 06 03:53:15 crc kubenswrapper[4980]: I1206 03:53:15.294930 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/92027099-00cc-45db-bbfe-7796fe5a68c4-log-httpd\") pod \"swift-proxy-85fcf7dd9-sn4m2\" (UID: \"92027099-00cc-45db-bbfe-7796fe5a68c4\") " pod="glance-kuttl-tests/swift-proxy-85fcf7dd9-sn4m2" Dec 06 03:53:15 crc kubenswrapper[4980]: I1206 03:53:15.395889 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/92027099-00cc-45db-bbfe-7796fe5a68c4-log-httpd\") pod \"swift-proxy-85fcf7dd9-sn4m2\" (UID: \"92027099-00cc-45db-bbfe-7796fe5a68c4\") " pod="glance-kuttl-tests/swift-proxy-85fcf7dd9-sn4m2" Dec 06 03:53:15 crc kubenswrapper[4980]: I1206 03:53:15.395958 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/92027099-00cc-45db-bbfe-7796fe5a68c4-etc-swift\") pod \"swift-proxy-85fcf7dd9-sn4m2\" (UID: \"92027099-00cc-45db-bbfe-7796fe5a68c4\") " pod="glance-kuttl-tests/swift-proxy-85fcf7dd9-sn4m2" Dec 06 03:53:15 crc kubenswrapper[4980]: I1206 03:53:15.396051 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/92027099-00cc-45db-bbfe-7796fe5a68c4-run-httpd\") pod \"swift-proxy-85fcf7dd9-sn4m2\" (UID: \"92027099-00cc-45db-bbfe-7796fe5a68c4\") " pod="glance-kuttl-tests/swift-proxy-85fcf7dd9-sn4m2" Dec 06 03:53:15 crc kubenswrapper[4980]: I1206 03:53:15.396083 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5c9mn\" (UniqueName: \"kubernetes.io/projected/92027099-00cc-45db-bbfe-7796fe5a68c4-kube-api-access-5c9mn\") pod \"swift-proxy-85fcf7dd9-sn4m2\" (UID: \"92027099-00cc-45db-bbfe-7796fe5a68c4\") " pod="glance-kuttl-tests/swift-proxy-85fcf7dd9-sn4m2" Dec 06 03:53:15 crc kubenswrapper[4980]: I1206 03:53:15.396109 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/92027099-00cc-45db-bbfe-7796fe5a68c4-config-data\") pod \"swift-proxy-85fcf7dd9-sn4m2\" (UID: \"92027099-00cc-45db-bbfe-7796fe5a68c4\") " pod="glance-kuttl-tests/swift-proxy-85fcf7dd9-sn4m2" Dec 06 03:53:15 crc kubenswrapper[4980]: I1206 03:53:15.396777 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/92027099-00cc-45db-bbfe-7796fe5a68c4-log-httpd\") pod \"swift-proxy-85fcf7dd9-sn4m2\" (UID: \"92027099-00cc-45db-bbfe-7796fe5a68c4\") " pod="glance-kuttl-tests/swift-proxy-85fcf7dd9-sn4m2" Dec 06 03:53:15 crc kubenswrapper[4980]: E1206 03:53:15.396874 4980 projected.go:288] Couldn't get configMap glance-kuttl-tests/swift-ring-files: configmap "swift-ring-files" not found Dec 06 03:53:15 crc kubenswrapper[4980]: E1206 03:53:15.396894 4980 projected.go:194] Error preparing data for projected volume etc-swift for pod glance-kuttl-tests/swift-proxy-85fcf7dd9-sn4m2: configmap "swift-ring-files" not found Dec 06 03:53:15 crc kubenswrapper[4980]: E1206 03:53:15.396962 4980 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/92027099-00cc-45db-bbfe-7796fe5a68c4-etc-swift podName:92027099-00cc-45db-bbfe-7796fe5a68c4 nodeName:}" failed. No retries permitted until 2025-12-06 03:53:15.896942158 +0000 UTC m=+1155.141693429 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/92027099-00cc-45db-bbfe-7796fe5a68c4-etc-swift") pod "swift-proxy-85fcf7dd9-sn4m2" (UID: "92027099-00cc-45db-bbfe-7796fe5a68c4") : configmap "swift-ring-files" not found Dec 06 03:53:15 crc kubenswrapper[4980]: I1206 03:53:15.397139 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/92027099-00cc-45db-bbfe-7796fe5a68c4-run-httpd\") pod \"swift-proxy-85fcf7dd9-sn4m2\" (UID: \"92027099-00cc-45db-bbfe-7796fe5a68c4\") " pod="glance-kuttl-tests/swift-proxy-85fcf7dd9-sn4m2" Dec 06 03:53:15 crc kubenswrapper[4980]: I1206 03:53:15.400653 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/92027099-00cc-45db-bbfe-7796fe5a68c4-config-data\") pod \"swift-proxy-85fcf7dd9-sn4m2\" (UID: \"92027099-00cc-45db-bbfe-7796fe5a68c4\") " pod="glance-kuttl-tests/swift-proxy-85fcf7dd9-sn4m2" Dec 06 03:53:15 crc kubenswrapper[4980]: I1206 03:53:15.439482 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5c9mn\" (UniqueName: \"kubernetes.io/projected/92027099-00cc-45db-bbfe-7796fe5a68c4-kube-api-access-5c9mn\") pod \"swift-proxy-85fcf7dd9-sn4m2\" (UID: \"92027099-00cc-45db-bbfe-7796fe5a68c4\") " pod="glance-kuttl-tests/swift-proxy-85fcf7dd9-sn4m2" Dec 06 03:53:15 crc kubenswrapper[4980]: I1206 03:53:15.903873 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/92027099-00cc-45db-bbfe-7796fe5a68c4-etc-swift\") pod \"swift-proxy-85fcf7dd9-sn4m2\" (UID: \"92027099-00cc-45db-bbfe-7796fe5a68c4\") " pod="glance-kuttl-tests/swift-proxy-85fcf7dd9-sn4m2" Dec 06 03:53:15 crc kubenswrapper[4980]: E1206 03:53:15.904418 4980 projected.go:288] Couldn't get configMap glance-kuttl-tests/swift-ring-files: configmap "swift-ring-files" not found Dec 06 03:53:15 crc kubenswrapper[4980]: E1206 03:53:15.904441 4980 projected.go:194] Error preparing data for projected volume etc-swift for pod glance-kuttl-tests/swift-proxy-85fcf7dd9-sn4m2: configmap "swift-ring-files" not found Dec 06 03:53:15 crc kubenswrapper[4980]: E1206 03:53:15.904493 4980 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/92027099-00cc-45db-bbfe-7796fe5a68c4-etc-swift podName:92027099-00cc-45db-bbfe-7796fe5a68c4 nodeName:}" failed. No retries permitted until 2025-12-06 03:53:16.904474937 +0000 UTC m=+1156.149226208 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/92027099-00cc-45db-bbfe-7796fe5a68c4-etc-swift") pod "swift-proxy-85fcf7dd9-sn4m2" (UID: "92027099-00cc-45db-bbfe-7796fe5a68c4") : configmap "swift-ring-files" not found Dec 06 03:53:16 crc kubenswrapper[4980]: I1206 03:53:16.073883 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-index-wt2kh" event={"ID":"79e40256-a186-48dc-b663-bd971d587fb9","Type":"ContainerStarted","Data":"af0ea6aa39d9f70811b560616f9ab6de8a276df24c5577bcf08826c305b33b3f"} Dec 06 03:53:16 crc kubenswrapper[4980]: I1206 03:53:16.074017 4980 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/glance-operator-index-wt2kh" podUID="79e40256-a186-48dc-b663-bd971d587fb9" containerName="registry-server" containerID="cri-o://af0ea6aa39d9f70811b560616f9ab6de8a276df24c5577bcf08826c305b33b3f" gracePeriod=2 Dec 06 03:53:16 crc kubenswrapper[4980]: I1206 03:53:16.095047 4980 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/glance-operator-index-wt2kh" podStartSLOduration=1.71827734 podStartE2EDuration="7.095019456s" podCreationTimestamp="2025-12-06 03:53:09 +0000 UTC" firstStartedPulling="2025-12-06 03:53:10.371584124 +0000 UTC m=+1149.616335395" lastFinishedPulling="2025-12-06 03:53:15.74832624 +0000 UTC m=+1154.993077511" observedRunningTime="2025-12-06 03:53:16.087279813 +0000 UTC m=+1155.332031084" watchObservedRunningTime="2025-12-06 03:53:16.095019456 +0000 UTC m=+1155.339770737" Dec 06 03:53:16 crc kubenswrapper[4980]: I1206 03:53:16.122074 4980 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-index-bzpqg"] Dec 06 03:53:16 crc kubenswrapper[4980]: W1206 03:53:16.177583 4980 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc88f8da4_819c_42bf_9938_60c898c6bace.slice/crio-735ffa39862ddd12438a5def7b1aa084724e689f23f792fb6ff103c384528c3f WatchSource:0}: Error finding container 735ffa39862ddd12438a5def7b1aa084724e689f23f792fb6ff103c384528c3f: Status 404 returned error can't find the container with id 735ffa39862ddd12438a5def7b1aa084724e689f23f792fb6ff103c384528c3f Dec 06 03:53:16 crc kubenswrapper[4980]: I1206 03:53:16.447973 4980 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-index-wt2kh" Dec 06 03:53:16 crc kubenswrapper[4980]: I1206 03:53:16.613366 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-txhbf\" (UniqueName: \"kubernetes.io/projected/79e40256-a186-48dc-b663-bd971d587fb9-kube-api-access-txhbf\") pod \"79e40256-a186-48dc-b663-bd971d587fb9\" (UID: \"79e40256-a186-48dc-b663-bd971d587fb9\") " Dec 06 03:53:16 crc kubenswrapper[4980]: I1206 03:53:16.618738 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/79e40256-a186-48dc-b663-bd971d587fb9-kube-api-access-txhbf" (OuterVolumeSpecName: "kube-api-access-txhbf") pod "79e40256-a186-48dc-b663-bd971d587fb9" (UID: "79e40256-a186-48dc-b663-bd971d587fb9"). InnerVolumeSpecName "kube-api-access-txhbf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 03:53:16 crc kubenswrapper[4980]: I1206 03:53:16.714639 4980 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-txhbf\" (UniqueName: \"kubernetes.io/projected/79e40256-a186-48dc-b663-bd971d587fb9-kube-api-access-txhbf\") on node \"crc\" DevicePath \"\"" Dec 06 03:53:16 crc kubenswrapper[4980]: I1206 03:53:16.917290 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/92027099-00cc-45db-bbfe-7796fe5a68c4-etc-swift\") pod \"swift-proxy-85fcf7dd9-sn4m2\" (UID: \"92027099-00cc-45db-bbfe-7796fe5a68c4\") " pod="glance-kuttl-tests/swift-proxy-85fcf7dd9-sn4m2" Dec 06 03:53:16 crc kubenswrapper[4980]: E1206 03:53:16.917535 4980 projected.go:288] Couldn't get configMap glance-kuttl-tests/swift-ring-files: configmap "swift-ring-files" not found Dec 06 03:53:16 crc kubenswrapper[4980]: E1206 03:53:16.917744 4980 projected.go:194] Error preparing data for projected volume etc-swift for pod glance-kuttl-tests/swift-proxy-85fcf7dd9-sn4m2: configmap "swift-ring-files" not found Dec 06 03:53:16 crc kubenswrapper[4980]: E1206 03:53:16.917812 4980 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/92027099-00cc-45db-bbfe-7796fe5a68c4-etc-swift podName:92027099-00cc-45db-bbfe-7796fe5a68c4 nodeName:}" failed. No retries permitted until 2025-12-06 03:53:18.917790066 +0000 UTC m=+1158.162541337 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/92027099-00cc-45db-bbfe-7796fe5a68c4-etc-swift") pod "swift-proxy-85fcf7dd9-sn4m2" (UID: "92027099-00cc-45db-bbfe-7796fe5a68c4") : configmap "swift-ring-files" not found Dec 06 03:53:17 crc kubenswrapper[4980]: I1206 03:53:17.084048 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-index-bzpqg" event={"ID":"c88f8da4-819c-42bf-9938-60c898c6bace","Type":"ContainerStarted","Data":"105516681feaff8d5c921cea0a01822025ffc27aedd0b89578c15cbb201837d8"} Dec 06 03:53:17 crc kubenswrapper[4980]: I1206 03:53:17.084096 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-index-bzpqg" event={"ID":"c88f8da4-819c-42bf-9938-60c898c6bace","Type":"ContainerStarted","Data":"735ffa39862ddd12438a5def7b1aa084724e689f23f792fb6ff103c384528c3f"} Dec 06 03:53:17 crc kubenswrapper[4980]: I1206 03:53:17.085821 4980 generic.go:334] "Generic (PLEG): container finished" podID="79e40256-a186-48dc-b663-bd971d587fb9" containerID="af0ea6aa39d9f70811b560616f9ab6de8a276df24c5577bcf08826c305b33b3f" exitCode=0 Dec 06 03:53:17 crc kubenswrapper[4980]: I1206 03:53:17.085889 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-index-wt2kh" event={"ID":"79e40256-a186-48dc-b663-bd971d587fb9","Type":"ContainerDied","Data":"af0ea6aa39d9f70811b560616f9ab6de8a276df24c5577bcf08826c305b33b3f"} Dec 06 03:53:17 crc kubenswrapper[4980]: I1206 03:53:17.085914 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-index-wt2kh" event={"ID":"79e40256-a186-48dc-b663-bd971d587fb9","Type":"ContainerDied","Data":"2fc9c0045a48fb08bec7a6d2aa972c88f65b3127b920d96dcde44d6583df9fdf"} Dec 06 03:53:17 crc kubenswrapper[4980]: I1206 03:53:17.085949 4980 scope.go:117] "RemoveContainer" containerID="af0ea6aa39d9f70811b560616f9ab6de8a276df24c5577bcf08826c305b33b3f" Dec 06 03:53:17 crc kubenswrapper[4980]: I1206 03:53:17.086081 4980 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-index-wt2kh" Dec 06 03:53:17 crc kubenswrapper[4980]: I1206 03:53:17.110505 4980 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/glance-operator-index-bzpqg" podStartSLOduration=3.062636039 podStartE2EDuration="3.110485467s" podCreationTimestamp="2025-12-06 03:53:14 +0000 UTC" firstStartedPulling="2025-12-06 03:53:16.182393483 +0000 UTC m=+1155.427144744" lastFinishedPulling="2025-12-06 03:53:16.230242901 +0000 UTC m=+1155.474994172" observedRunningTime="2025-12-06 03:53:17.104154714 +0000 UTC m=+1156.348906005" watchObservedRunningTime="2025-12-06 03:53:17.110485467 +0000 UTC m=+1156.355236748" Dec 06 03:53:17 crc kubenswrapper[4980]: I1206 03:53:17.122856 4980 scope.go:117] "RemoveContainer" containerID="af0ea6aa39d9f70811b560616f9ab6de8a276df24c5577bcf08826c305b33b3f" Dec 06 03:53:17 crc kubenswrapper[4980]: I1206 03:53:17.123168 4980 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/glance-operator-index-wt2kh"] Dec 06 03:53:17 crc kubenswrapper[4980]: E1206 03:53:17.123641 4980 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"af0ea6aa39d9f70811b560616f9ab6de8a276df24c5577bcf08826c305b33b3f\": container with ID starting with af0ea6aa39d9f70811b560616f9ab6de8a276df24c5577bcf08826c305b33b3f not found: ID does not exist" containerID="af0ea6aa39d9f70811b560616f9ab6de8a276df24c5577bcf08826c305b33b3f" Dec 06 03:53:17 crc kubenswrapper[4980]: I1206 03:53:17.123678 4980 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"af0ea6aa39d9f70811b560616f9ab6de8a276df24c5577bcf08826c305b33b3f"} err="failed to get container status \"af0ea6aa39d9f70811b560616f9ab6de8a276df24c5577bcf08826c305b33b3f\": rpc error: code = NotFound desc = could not find container \"af0ea6aa39d9f70811b560616f9ab6de8a276df24c5577bcf08826c305b33b3f\": container with ID starting with af0ea6aa39d9f70811b560616f9ab6de8a276df24c5577bcf08826c305b33b3f not found: ID does not exist" Dec 06 03:53:17 crc kubenswrapper[4980]: I1206 03:53:17.131995 4980 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/glance-operator-index-wt2kh"] Dec 06 03:53:17 crc kubenswrapper[4980]: I1206 03:53:17.195474 4980 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="79e40256-a186-48dc-b663-bd971d587fb9" path="/var/lib/kubelet/pods/79e40256-a186-48dc-b663-bd971d587fb9/volumes" Dec 06 03:53:18 crc kubenswrapper[4980]: I1206 03:53:18.370929 4980 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/swift-storage-0"] Dec 06 03:53:18 crc kubenswrapper[4980]: E1206 03:53:18.371733 4980 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="79e40256-a186-48dc-b663-bd971d587fb9" containerName="registry-server" Dec 06 03:53:18 crc kubenswrapper[4980]: I1206 03:53:18.371756 4980 state_mem.go:107] "Deleted CPUSet assignment" podUID="79e40256-a186-48dc-b663-bd971d587fb9" containerName="registry-server" Dec 06 03:53:18 crc kubenswrapper[4980]: I1206 03:53:18.371949 4980 memory_manager.go:354] "RemoveStaleState removing state" podUID="79e40256-a186-48dc-b663-bd971d587fb9" containerName="registry-server" Dec 06 03:53:18 crc kubenswrapper[4980]: I1206 03:53:18.377849 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/swift-storage-0" Dec 06 03:53:18 crc kubenswrapper[4980]: I1206 03:53:18.381748 4980 reflector.go:368] Caches populated for *v1.ConfigMap from object-"glance-kuttl-tests"/"swift-storage-config-data" Dec 06 03:53:18 crc kubenswrapper[4980]: I1206 03:53:18.402949 4980 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/swift-storage-0"] Dec 06 03:53:18 crc kubenswrapper[4980]: I1206 03:53:18.551821 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/3752191f-1260-4be2-8d68-308491d20468-cache\") pod \"swift-storage-0\" (UID: \"3752191f-1260-4be2-8d68-308491d20468\") " pod="glance-kuttl-tests/swift-storage-0" Dec 06 03:53:18 crc kubenswrapper[4980]: I1206 03:53:18.551892 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"swift-storage-0\" (UID: \"3752191f-1260-4be2-8d68-308491d20468\") " pod="glance-kuttl-tests/swift-storage-0" Dec 06 03:53:18 crc kubenswrapper[4980]: I1206 03:53:18.552005 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rfnt8\" (UniqueName: \"kubernetes.io/projected/3752191f-1260-4be2-8d68-308491d20468-kube-api-access-rfnt8\") pod \"swift-storage-0\" (UID: \"3752191f-1260-4be2-8d68-308491d20468\") " pod="glance-kuttl-tests/swift-storage-0" Dec 06 03:53:18 crc kubenswrapper[4980]: I1206 03:53:18.552037 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/3752191f-1260-4be2-8d68-308491d20468-lock\") pod \"swift-storage-0\" (UID: \"3752191f-1260-4be2-8d68-308491d20468\") " pod="glance-kuttl-tests/swift-storage-0" Dec 06 03:53:18 crc kubenswrapper[4980]: I1206 03:53:18.552101 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/3752191f-1260-4be2-8d68-308491d20468-etc-swift\") pod \"swift-storage-0\" (UID: \"3752191f-1260-4be2-8d68-308491d20468\") " pod="glance-kuttl-tests/swift-storage-0" Dec 06 03:53:18 crc kubenswrapper[4980]: I1206 03:53:18.654260 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/3752191f-1260-4be2-8d68-308491d20468-etc-swift\") pod \"swift-storage-0\" (UID: \"3752191f-1260-4be2-8d68-308491d20468\") " pod="glance-kuttl-tests/swift-storage-0" Dec 06 03:53:18 crc kubenswrapper[4980]: I1206 03:53:18.654402 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/3752191f-1260-4be2-8d68-308491d20468-cache\") pod \"swift-storage-0\" (UID: \"3752191f-1260-4be2-8d68-308491d20468\") " pod="glance-kuttl-tests/swift-storage-0" Dec 06 03:53:18 crc kubenswrapper[4980]: I1206 03:53:18.654452 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"swift-storage-0\" (UID: \"3752191f-1260-4be2-8d68-308491d20468\") " pod="glance-kuttl-tests/swift-storage-0" Dec 06 03:53:18 crc kubenswrapper[4980]: E1206 03:53:18.654539 4980 projected.go:288] Couldn't get configMap glance-kuttl-tests/swift-ring-files: configmap "swift-ring-files" not found Dec 06 03:53:18 crc kubenswrapper[4980]: E1206 03:53:18.654572 4980 projected.go:194] Error preparing data for projected volume etc-swift for pod glance-kuttl-tests/swift-storage-0: configmap "swift-ring-files" not found Dec 06 03:53:18 crc kubenswrapper[4980]: E1206 03:53:18.654648 4980 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3752191f-1260-4be2-8d68-308491d20468-etc-swift podName:3752191f-1260-4be2-8d68-308491d20468 nodeName:}" failed. No retries permitted until 2025-12-06 03:53:19.154612404 +0000 UTC m=+1158.399363675 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/3752191f-1260-4be2-8d68-308491d20468-etc-swift") pod "swift-storage-0" (UID: "3752191f-1260-4be2-8d68-308491d20468") : configmap "swift-ring-files" not found Dec 06 03:53:18 crc kubenswrapper[4980]: I1206 03:53:18.654696 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rfnt8\" (UniqueName: \"kubernetes.io/projected/3752191f-1260-4be2-8d68-308491d20468-kube-api-access-rfnt8\") pod \"swift-storage-0\" (UID: \"3752191f-1260-4be2-8d68-308491d20468\") " pod="glance-kuttl-tests/swift-storage-0" Dec 06 03:53:18 crc kubenswrapper[4980]: I1206 03:53:18.654759 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/3752191f-1260-4be2-8d68-308491d20468-lock\") pod \"swift-storage-0\" (UID: \"3752191f-1260-4be2-8d68-308491d20468\") " pod="glance-kuttl-tests/swift-storage-0" Dec 06 03:53:18 crc kubenswrapper[4980]: I1206 03:53:18.655037 4980 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"swift-storage-0\" (UID: \"3752191f-1260-4be2-8d68-308491d20468\") device mount path \"/mnt/openstack/pv06\"" pod="glance-kuttl-tests/swift-storage-0" Dec 06 03:53:18 crc kubenswrapper[4980]: I1206 03:53:18.655424 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/3752191f-1260-4be2-8d68-308491d20468-cache\") pod \"swift-storage-0\" (UID: \"3752191f-1260-4be2-8d68-308491d20468\") " pod="glance-kuttl-tests/swift-storage-0" Dec 06 03:53:18 crc kubenswrapper[4980]: I1206 03:53:18.655783 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/3752191f-1260-4be2-8d68-308491d20468-lock\") pod \"swift-storage-0\" (UID: \"3752191f-1260-4be2-8d68-308491d20468\") " pod="glance-kuttl-tests/swift-storage-0" Dec 06 03:53:18 crc kubenswrapper[4980]: I1206 03:53:18.676877 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rfnt8\" (UniqueName: \"kubernetes.io/projected/3752191f-1260-4be2-8d68-308491d20468-kube-api-access-rfnt8\") pod \"swift-storage-0\" (UID: \"3752191f-1260-4be2-8d68-308491d20468\") " pod="glance-kuttl-tests/swift-storage-0" Dec 06 03:53:18 crc kubenswrapper[4980]: I1206 03:53:18.682810 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"swift-storage-0\" (UID: \"3752191f-1260-4be2-8d68-308491d20468\") " pod="glance-kuttl-tests/swift-storage-0" Dec 06 03:53:18 crc kubenswrapper[4980]: I1206 03:53:18.960045 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/92027099-00cc-45db-bbfe-7796fe5a68c4-etc-swift\") pod \"swift-proxy-85fcf7dd9-sn4m2\" (UID: \"92027099-00cc-45db-bbfe-7796fe5a68c4\") " pod="glance-kuttl-tests/swift-proxy-85fcf7dd9-sn4m2" Dec 06 03:53:18 crc kubenswrapper[4980]: E1206 03:53:18.960264 4980 projected.go:288] Couldn't get configMap glance-kuttl-tests/swift-ring-files: configmap "swift-ring-files" not found Dec 06 03:53:18 crc kubenswrapper[4980]: E1206 03:53:18.960306 4980 projected.go:194] Error preparing data for projected volume etc-swift for pod glance-kuttl-tests/swift-proxy-85fcf7dd9-sn4m2: configmap "swift-ring-files" not found Dec 06 03:53:18 crc kubenswrapper[4980]: E1206 03:53:18.960387 4980 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/92027099-00cc-45db-bbfe-7796fe5a68c4-etc-swift podName:92027099-00cc-45db-bbfe-7796fe5a68c4 nodeName:}" failed. No retries permitted until 2025-12-06 03:53:22.960359881 +0000 UTC m=+1162.205111182 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/92027099-00cc-45db-bbfe-7796fe5a68c4-etc-swift") pod "swift-proxy-85fcf7dd9-sn4m2" (UID: "92027099-00cc-45db-bbfe-7796fe5a68c4") : configmap "swift-ring-files" not found Dec 06 03:53:19 crc kubenswrapper[4980]: I1206 03:53:19.163074 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/3752191f-1260-4be2-8d68-308491d20468-etc-swift\") pod \"swift-storage-0\" (UID: \"3752191f-1260-4be2-8d68-308491d20468\") " pod="glance-kuttl-tests/swift-storage-0" Dec 06 03:53:19 crc kubenswrapper[4980]: E1206 03:53:19.163313 4980 projected.go:288] Couldn't get configMap glance-kuttl-tests/swift-ring-files: configmap "swift-ring-files" not found Dec 06 03:53:19 crc kubenswrapper[4980]: E1206 03:53:19.163567 4980 projected.go:194] Error preparing data for projected volume etc-swift for pod glance-kuttl-tests/swift-storage-0: configmap "swift-ring-files" not found Dec 06 03:53:19 crc kubenswrapper[4980]: E1206 03:53:19.163641 4980 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3752191f-1260-4be2-8d68-308491d20468-etc-swift podName:3752191f-1260-4be2-8d68-308491d20468 nodeName:}" failed. No retries permitted until 2025-12-06 03:53:20.163616266 +0000 UTC m=+1159.408367547 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/3752191f-1260-4be2-8d68-308491d20468-etc-swift") pod "swift-storage-0" (UID: "3752191f-1260-4be2-8d68-308491d20468") : configmap "swift-ring-files" not found Dec 06 03:53:20 crc kubenswrapper[4980]: I1206 03:53:20.176674 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/3752191f-1260-4be2-8d68-308491d20468-etc-swift\") pod \"swift-storage-0\" (UID: \"3752191f-1260-4be2-8d68-308491d20468\") " pod="glance-kuttl-tests/swift-storage-0" Dec 06 03:53:20 crc kubenswrapper[4980]: E1206 03:53:20.176995 4980 projected.go:288] Couldn't get configMap glance-kuttl-tests/swift-ring-files: configmap "swift-ring-files" not found Dec 06 03:53:20 crc kubenswrapper[4980]: E1206 03:53:20.177008 4980 projected.go:194] Error preparing data for projected volume etc-swift for pod glance-kuttl-tests/swift-storage-0: configmap "swift-ring-files" not found Dec 06 03:53:20 crc kubenswrapper[4980]: E1206 03:53:20.177049 4980 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3752191f-1260-4be2-8d68-308491d20468-etc-swift podName:3752191f-1260-4be2-8d68-308491d20468 nodeName:}" failed. No retries permitted until 2025-12-06 03:53:22.177036137 +0000 UTC m=+1161.421787408 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/3752191f-1260-4be2-8d68-308491d20468-etc-swift") pod "swift-storage-0" (UID: "3752191f-1260-4be2-8d68-308491d20468") : configmap "swift-ring-files" not found Dec 06 03:53:21 crc kubenswrapper[4980]: I1206 03:53:21.896294 4980 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/swift-ring-rebalance-vn7bt"] Dec 06 03:53:21 crc kubenswrapper[4980]: I1206 03:53:21.897751 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/swift-ring-rebalance-vn7bt" Dec 06 03:53:21 crc kubenswrapper[4980]: I1206 03:53:21.901721 4980 reflector.go:368] Caches populated for *v1.ConfigMap from object-"glance-kuttl-tests"/"swift-ring-config-data" Dec 06 03:53:21 crc kubenswrapper[4980]: I1206 03:53:21.901747 4980 reflector.go:368] Caches populated for *v1.ConfigMap from object-"glance-kuttl-tests"/"swift-ring-scripts" Dec 06 03:53:21 crc kubenswrapper[4980]: I1206 03:53:21.915357 4980 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/swift-ring-rebalance-vn7bt"] Dec 06 03:53:21 crc kubenswrapper[4980]: E1206 03:53:21.916640 4980 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[dispersionconf etc-swift kube-api-access-bqmd8 ring-data-devices scripts swiftconf], unattached volumes=[], failed to process volumes=[]: context canceled" pod="glance-kuttl-tests/swift-ring-rebalance-vn7bt" podUID="b9146130-53f7-4c79-adff-c093da73c967" Dec 06 03:53:21 crc kubenswrapper[4980]: I1206 03:53:21.925640 4980 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/swift-ring-rebalance-pfjgb"] Dec 06 03:53:21 crc kubenswrapper[4980]: I1206 03:53:21.926866 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/swift-ring-rebalance-pfjgb" Dec 06 03:53:21 crc kubenswrapper[4980]: I1206 03:53:21.938898 4980 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/swift-ring-rebalance-pfjgb"] Dec 06 03:53:21 crc kubenswrapper[4980]: I1206 03:53:21.944991 4980 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/swift-ring-rebalance-vn7bt"] Dec 06 03:53:22 crc kubenswrapper[4980]: I1206 03:53:22.003807 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/b9146130-53f7-4c79-adff-c093da73c967-swiftconf\") pod \"swift-ring-rebalance-vn7bt\" (UID: \"b9146130-53f7-4c79-adff-c093da73c967\") " pod="glance-kuttl-tests/swift-ring-rebalance-vn7bt" Dec 06 03:53:22 crc kubenswrapper[4980]: I1206 03:53:22.003866 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/b9146130-53f7-4c79-adff-c093da73c967-ring-data-devices\") pod \"swift-ring-rebalance-vn7bt\" (UID: \"b9146130-53f7-4c79-adff-c093da73c967\") " pod="glance-kuttl-tests/swift-ring-rebalance-vn7bt" Dec 06 03:53:22 crc kubenswrapper[4980]: I1206 03:53:22.003899 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b9146130-53f7-4c79-adff-c093da73c967-scripts\") pod \"swift-ring-rebalance-vn7bt\" (UID: \"b9146130-53f7-4c79-adff-c093da73c967\") " pod="glance-kuttl-tests/swift-ring-rebalance-vn7bt" Dec 06 03:53:22 crc kubenswrapper[4980]: I1206 03:53:22.003925 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/b9146130-53f7-4c79-adff-c093da73c967-etc-swift\") pod \"swift-ring-rebalance-vn7bt\" (UID: \"b9146130-53f7-4c79-adff-c093da73c967\") " pod="glance-kuttl-tests/swift-ring-rebalance-vn7bt" Dec 06 03:53:22 crc kubenswrapper[4980]: I1206 03:53:22.004015 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/b9146130-53f7-4c79-adff-c093da73c967-dispersionconf\") pod \"swift-ring-rebalance-vn7bt\" (UID: \"b9146130-53f7-4c79-adff-c093da73c967\") " pod="glance-kuttl-tests/swift-ring-rebalance-vn7bt" Dec 06 03:53:22 crc kubenswrapper[4980]: I1206 03:53:22.004265 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bqmd8\" (UniqueName: \"kubernetes.io/projected/b9146130-53f7-4c79-adff-c093da73c967-kube-api-access-bqmd8\") pod \"swift-ring-rebalance-vn7bt\" (UID: \"b9146130-53f7-4c79-adff-c093da73c967\") " pod="glance-kuttl-tests/swift-ring-rebalance-vn7bt" Dec 06 03:53:22 crc kubenswrapper[4980]: I1206 03:53:22.105182 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/b9146130-53f7-4c79-adff-c093da73c967-dispersionconf\") pod \"swift-ring-rebalance-vn7bt\" (UID: \"b9146130-53f7-4c79-adff-c093da73c967\") " pod="glance-kuttl-tests/swift-ring-rebalance-vn7bt" Dec 06 03:53:22 crc kubenswrapper[4980]: I1206 03:53:22.105240 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/43cddc9e-ad31-4596-8e89-cb3260fd55a8-etc-swift\") pod \"swift-ring-rebalance-pfjgb\" (UID: \"43cddc9e-ad31-4596-8e89-cb3260fd55a8\") " pod="glance-kuttl-tests/swift-ring-rebalance-pfjgb" Dec 06 03:53:22 crc kubenswrapper[4980]: I1206 03:53:22.105265 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/43cddc9e-ad31-4596-8e89-cb3260fd55a8-ring-data-devices\") pod \"swift-ring-rebalance-pfjgb\" (UID: \"43cddc9e-ad31-4596-8e89-cb3260fd55a8\") " pod="glance-kuttl-tests/swift-ring-rebalance-pfjgb" Dec 06 03:53:22 crc kubenswrapper[4980]: I1206 03:53:22.105281 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/43cddc9e-ad31-4596-8e89-cb3260fd55a8-dispersionconf\") pod \"swift-ring-rebalance-pfjgb\" (UID: \"43cddc9e-ad31-4596-8e89-cb3260fd55a8\") " pod="glance-kuttl-tests/swift-ring-rebalance-pfjgb" Dec 06 03:53:22 crc kubenswrapper[4980]: I1206 03:53:22.105329 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/43cddc9e-ad31-4596-8e89-cb3260fd55a8-swiftconf\") pod \"swift-ring-rebalance-pfjgb\" (UID: \"43cddc9e-ad31-4596-8e89-cb3260fd55a8\") " pod="glance-kuttl-tests/swift-ring-rebalance-pfjgb" Dec 06 03:53:22 crc kubenswrapper[4980]: I1206 03:53:22.105350 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bqmd8\" (UniqueName: \"kubernetes.io/projected/b9146130-53f7-4c79-adff-c093da73c967-kube-api-access-bqmd8\") pod \"swift-ring-rebalance-vn7bt\" (UID: \"b9146130-53f7-4c79-adff-c093da73c967\") " pod="glance-kuttl-tests/swift-ring-rebalance-vn7bt" Dec 06 03:53:22 crc kubenswrapper[4980]: I1206 03:53:22.105482 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/b9146130-53f7-4c79-adff-c093da73c967-swiftconf\") pod \"swift-ring-rebalance-vn7bt\" (UID: \"b9146130-53f7-4c79-adff-c093da73c967\") " pod="glance-kuttl-tests/swift-ring-rebalance-vn7bt" Dec 06 03:53:22 crc kubenswrapper[4980]: I1206 03:53:22.105575 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/b9146130-53f7-4c79-adff-c093da73c967-ring-data-devices\") pod \"swift-ring-rebalance-vn7bt\" (UID: \"b9146130-53f7-4c79-adff-c093da73c967\") " pod="glance-kuttl-tests/swift-ring-rebalance-vn7bt" Dec 06 03:53:22 crc kubenswrapper[4980]: I1206 03:53:22.105610 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b9146130-53f7-4c79-adff-c093da73c967-scripts\") pod \"swift-ring-rebalance-vn7bt\" (UID: \"b9146130-53f7-4c79-adff-c093da73c967\") " pod="glance-kuttl-tests/swift-ring-rebalance-vn7bt" Dec 06 03:53:22 crc kubenswrapper[4980]: I1206 03:53:22.105636 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/b9146130-53f7-4c79-adff-c093da73c967-etc-swift\") pod \"swift-ring-rebalance-vn7bt\" (UID: \"b9146130-53f7-4c79-adff-c093da73c967\") " pod="glance-kuttl-tests/swift-ring-rebalance-vn7bt" Dec 06 03:53:22 crc kubenswrapper[4980]: I1206 03:53:22.105658 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/43cddc9e-ad31-4596-8e89-cb3260fd55a8-scripts\") pod \"swift-ring-rebalance-pfjgb\" (UID: \"43cddc9e-ad31-4596-8e89-cb3260fd55a8\") " pod="glance-kuttl-tests/swift-ring-rebalance-pfjgb" Dec 06 03:53:22 crc kubenswrapper[4980]: I1206 03:53:22.105699 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hdpzd\" (UniqueName: \"kubernetes.io/projected/43cddc9e-ad31-4596-8e89-cb3260fd55a8-kube-api-access-hdpzd\") pod \"swift-ring-rebalance-pfjgb\" (UID: \"43cddc9e-ad31-4596-8e89-cb3260fd55a8\") " pod="glance-kuttl-tests/swift-ring-rebalance-pfjgb" Dec 06 03:53:22 crc kubenswrapper[4980]: I1206 03:53:22.107788 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/b9146130-53f7-4c79-adff-c093da73c967-etc-swift\") pod \"swift-ring-rebalance-vn7bt\" (UID: \"b9146130-53f7-4c79-adff-c093da73c967\") " pod="glance-kuttl-tests/swift-ring-rebalance-vn7bt" Dec 06 03:53:22 crc kubenswrapper[4980]: I1206 03:53:22.108170 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/b9146130-53f7-4c79-adff-c093da73c967-ring-data-devices\") pod \"swift-ring-rebalance-vn7bt\" (UID: \"b9146130-53f7-4c79-adff-c093da73c967\") " pod="glance-kuttl-tests/swift-ring-rebalance-vn7bt" Dec 06 03:53:22 crc kubenswrapper[4980]: I1206 03:53:22.108547 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b9146130-53f7-4c79-adff-c093da73c967-scripts\") pod \"swift-ring-rebalance-vn7bt\" (UID: \"b9146130-53f7-4c79-adff-c093da73c967\") " pod="glance-kuttl-tests/swift-ring-rebalance-vn7bt" Dec 06 03:53:22 crc kubenswrapper[4980]: I1206 03:53:22.111387 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/b9146130-53f7-4c79-adff-c093da73c967-swiftconf\") pod \"swift-ring-rebalance-vn7bt\" (UID: \"b9146130-53f7-4c79-adff-c093da73c967\") " pod="glance-kuttl-tests/swift-ring-rebalance-vn7bt" Dec 06 03:53:22 crc kubenswrapper[4980]: I1206 03:53:22.119832 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/b9146130-53f7-4c79-adff-c093da73c967-dispersionconf\") pod \"swift-ring-rebalance-vn7bt\" (UID: \"b9146130-53f7-4c79-adff-c093da73c967\") " pod="glance-kuttl-tests/swift-ring-rebalance-vn7bt" Dec 06 03:53:22 crc kubenswrapper[4980]: I1206 03:53:22.132480 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/swift-ring-rebalance-vn7bt" Dec 06 03:53:22 crc kubenswrapper[4980]: I1206 03:53:22.133751 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bqmd8\" (UniqueName: \"kubernetes.io/projected/b9146130-53f7-4c79-adff-c093da73c967-kube-api-access-bqmd8\") pod \"swift-ring-rebalance-vn7bt\" (UID: \"b9146130-53f7-4c79-adff-c093da73c967\") " pod="glance-kuttl-tests/swift-ring-rebalance-vn7bt" Dec 06 03:53:22 crc kubenswrapper[4980]: I1206 03:53:22.182647 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/swift-ring-rebalance-vn7bt" Dec 06 03:53:22 crc kubenswrapper[4980]: I1206 03:53:22.208866 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/3752191f-1260-4be2-8d68-308491d20468-etc-swift\") pod \"swift-storage-0\" (UID: \"3752191f-1260-4be2-8d68-308491d20468\") " pod="glance-kuttl-tests/swift-storage-0" Dec 06 03:53:22 crc kubenswrapper[4980]: I1206 03:53:22.208941 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/43cddc9e-ad31-4596-8e89-cb3260fd55a8-etc-swift\") pod \"swift-ring-rebalance-pfjgb\" (UID: \"43cddc9e-ad31-4596-8e89-cb3260fd55a8\") " pod="glance-kuttl-tests/swift-ring-rebalance-pfjgb" Dec 06 03:53:22 crc kubenswrapper[4980]: I1206 03:53:22.208977 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/43cddc9e-ad31-4596-8e89-cb3260fd55a8-ring-data-devices\") pod \"swift-ring-rebalance-pfjgb\" (UID: \"43cddc9e-ad31-4596-8e89-cb3260fd55a8\") " pod="glance-kuttl-tests/swift-ring-rebalance-pfjgb" Dec 06 03:53:22 crc kubenswrapper[4980]: I1206 03:53:22.208997 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/43cddc9e-ad31-4596-8e89-cb3260fd55a8-dispersionconf\") pod \"swift-ring-rebalance-pfjgb\" (UID: \"43cddc9e-ad31-4596-8e89-cb3260fd55a8\") " pod="glance-kuttl-tests/swift-ring-rebalance-pfjgb" Dec 06 03:53:22 crc kubenswrapper[4980]: I1206 03:53:22.209056 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/43cddc9e-ad31-4596-8e89-cb3260fd55a8-swiftconf\") pod \"swift-ring-rebalance-pfjgb\" (UID: \"43cddc9e-ad31-4596-8e89-cb3260fd55a8\") " pod="glance-kuttl-tests/swift-ring-rebalance-pfjgb" Dec 06 03:53:22 crc kubenswrapper[4980]: I1206 03:53:22.209128 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/43cddc9e-ad31-4596-8e89-cb3260fd55a8-scripts\") pod \"swift-ring-rebalance-pfjgb\" (UID: \"43cddc9e-ad31-4596-8e89-cb3260fd55a8\") " pod="glance-kuttl-tests/swift-ring-rebalance-pfjgb" Dec 06 03:53:22 crc kubenswrapper[4980]: I1206 03:53:22.209153 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hdpzd\" (UniqueName: \"kubernetes.io/projected/43cddc9e-ad31-4596-8e89-cb3260fd55a8-kube-api-access-hdpzd\") pod \"swift-ring-rebalance-pfjgb\" (UID: \"43cddc9e-ad31-4596-8e89-cb3260fd55a8\") " pod="glance-kuttl-tests/swift-ring-rebalance-pfjgb" Dec 06 03:53:22 crc kubenswrapper[4980]: E1206 03:53:22.209586 4980 projected.go:288] Couldn't get configMap glance-kuttl-tests/swift-ring-files: configmap "swift-ring-files" not found Dec 06 03:53:22 crc kubenswrapper[4980]: E1206 03:53:22.209608 4980 projected.go:194] Error preparing data for projected volume etc-swift for pod glance-kuttl-tests/swift-storage-0: configmap "swift-ring-files" not found Dec 06 03:53:22 crc kubenswrapper[4980]: E1206 03:53:22.209659 4980 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3752191f-1260-4be2-8d68-308491d20468-etc-swift podName:3752191f-1260-4be2-8d68-308491d20468 nodeName:}" failed. No retries permitted until 2025-12-06 03:53:26.209641465 +0000 UTC m=+1165.454392746 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/3752191f-1260-4be2-8d68-308491d20468-etc-swift") pod "swift-storage-0" (UID: "3752191f-1260-4be2-8d68-308491d20468") : configmap "swift-ring-files" not found Dec 06 03:53:22 crc kubenswrapper[4980]: I1206 03:53:22.210606 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/43cddc9e-ad31-4596-8e89-cb3260fd55a8-scripts\") pod \"swift-ring-rebalance-pfjgb\" (UID: \"43cddc9e-ad31-4596-8e89-cb3260fd55a8\") " pod="glance-kuttl-tests/swift-ring-rebalance-pfjgb" Dec 06 03:53:22 crc kubenswrapper[4980]: I1206 03:53:22.210961 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/43cddc9e-ad31-4596-8e89-cb3260fd55a8-etc-swift\") pod \"swift-ring-rebalance-pfjgb\" (UID: \"43cddc9e-ad31-4596-8e89-cb3260fd55a8\") " pod="glance-kuttl-tests/swift-ring-rebalance-pfjgb" Dec 06 03:53:22 crc kubenswrapper[4980]: I1206 03:53:22.211346 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/43cddc9e-ad31-4596-8e89-cb3260fd55a8-ring-data-devices\") pod \"swift-ring-rebalance-pfjgb\" (UID: \"43cddc9e-ad31-4596-8e89-cb3260fd55a8\") " pod="glance-kuttl-tests/swift-ring-rebalance-pfjgb" Dec 06 03:53:22 crc kubenswrapper[4980]: I1206 03:53:22.215066 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/43cddc9e-ad31-4596-8e89-cb3260fd55a8-dispersionconf\") pod \"swift-ring-rebalance-pfjgb\" (UID: \"43cddc9e-ad31-4596-8e89-cb3260fd55a8\") " pod="glance-kuttl-tests/swift-ring-rebalance-pfjgb" Dec 06 03:53:22 crc kubenswrapper[4980]: I1206 03:53:22.219289 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/43cddc9e-ad31-4596-8e89-cb3260fd55a8-swiftconf\") pod \"swift-ring-rebalance-pfjgb\" (UID: \"43cddc9e-ad31-4596-8e89-cb3260fd55a8\") " pod="glance-kuttl-tests/swift-ring-rebalance-pfjgb" Dec 06 03:53:22 crc kubenswrapper[4980]: I1206 03:53:22.247280 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hdpzd\" (UniqueName: \"kubernetes.io/projected/43cddc9e-ad31-4596-8e89-cb3260fd55a8-kube-api-access-hdpzd\") pod \"swift-ring-rebalance-pfjgb\" (UID: \"43cddc9e-ad31-4596-8e89-cb3260fd55a8\") " pod="glance-kuttl-tests/swift-ring-rebalance-pfjgb" Dec 06 03:53:22 crc kubenswrapper[4980]: I1206 03:53:22.326304 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b9146130-53f7-4c79-adff-c093da73c967-scripts\") pod \"b9146130-53f7-4c79-adff-c093da73c967\" (UID: \"b9146130-53f7-4c79-adff-c093da73c967\") " Dec 06 03:53:22 crc kubenswrapper[4980]: I1206 03:53:22.326449 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/b9146130-53f7-4c79-adff-c093da73c967-ring-data-devices\") pod \"b9146130-53f7-4c79-adff-c093da73c967\" (UID: \"b9146130-53f7-4c79-adff-c093da73c967\") " Dec 06 03:53:22 crc kubenswrapper[4980]: I1206 03:53:22.326533 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/b9146130-53f7-4c79-adff-c093da73c967-dispersionconf\") pod \"b9146130-53f7-4c79-adff-c093da73c967\" (UID: \"b9146130-53f7-4c79-adff-c093da73c967\") " Dec 06 03:53:22 crc kubenswrapper[4980]: I1206 03:53:22.326613 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/b9146130-53f7-4c79-adff-c093da73c967-swiftconf\") pod \"b9146130-53f7-4c79-adff-c093da73c967\" (UID: \"b9146130-53f7-4c79-adff-c093da73c967\") " Dec 06 03:53:22 crc kubenswrapper[4980]: I1206 03:53:22.326645 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bqmd8\" (UniqueName: \"kubernetes.io/projected/b9146130-53f7-4c79-adff-c093da73c967-kube-api-access-bqmd8\") pod \"b9146130-53f7-4c79-adff-c093da73c967\" (UID: \"b9146130-53f7-4c79-adff-c093da73c967\") " Dec 06 03:53:22 crc kubenswrapper[4980]: I1206 03:53:22.326683 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/b9146130-53f7-4c79-adff-c093da73c967-etc-swift\") pod \"b9146130-53f7-4c79-adff-c093da73c967\" (UID: \"b9146130-53f7-4c79-adff-c093da73c967\") " Dec 06 03:53:22 crc kubenswrapper[4980]: I1206 03:53:22.327399 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b9146130-53f7-4c79-adff-c093da73c967-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "b9146130-53f7-4c79-adff-c093da73c967" (UID: "b9146130-53f7-4c79-adff-c093da73c967"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 03:53:22 crc kubenswrapper[4980]: I1206 03:53:22.327768 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b9146130-53f7-4c79-adff-c093da73c967-scripts" (OuterVolumeSpecName: "scripts") pod "b9146130-53f7-4c79-adff-c093da73c967" (UID: "b9146130-53f7-4c79-adff-c093da73c967"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 03:53:22 crc kubenswrapper[4980]: I1206 03:53:22.328053 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b9146130-53f7-4c79-adff-c093da73c967-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "b9146130-53f7-4c79-adff-c093da73c967" (UID: "b9146130-53f7-4c79-adff-c093da73c967"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 03:53:22 crc kubenswrapper[4980]: I1206 03:53:22.332559 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b9146130-53f7-4c79-adff-c093da73c967-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "b9146130-53f7-4c79-adff-c093da73c967" (UID: "b9146130-53f7-4c79-adff-c093da73c967"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 03:53:22 crc kubenswrapper[4980]: I1206 03:53:22.332725 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b9146130-53f7-4c79-adff-c093da73c967-kube-api-access-bqmd8" (OuterVolumeSpecName: "kube-api-access-bqmd8") pod "b9146130-53f7-4c79-adff-c093da73c967" (UID: "b9146130-53f7-4c79-adff-c093da73c967"). InnerVolumeSpecName "kube-api-access-bqmd8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 03:53:22 crc kubenswrapper[4980]: I1206 03:53:22.333353 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b9146130-53f7-4c79-adff-c093da73c967-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "b9146130-53f7-4c79-adff-c093da73c967" (UID: "b9146130-53f7-4c79-adff-c093da73c967"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 03:53:22 crc kubenswrapper[4980]: I1206 03:53:22.428886 4980 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/b9146130-53f7-4c79-adff-c093da73c967-ring-data-devices\") on node \"crc\" DevicePath \"\"" Dec 06 03:53:22 crc kubenswrapper[4980]: I1206 03:53:22.428940 4980 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/b9146130-53f7-4c79-adff-c093da73c967-dispersionconf\") on node \"crc\" DevicePath \"\"" Dec 06 03:53:22 crc kubenswrapper[4980]: I1206 03:53:22.428958 4980 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/b9146130-53f7-4c79-adff-c093da73c967-swiftconf\") on node \"crc\" DevicePath \"\"" Dec 06 03:53:22 crc kubenswrapper[4980]: I1206 03:53:22.428976 4980 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bqmd8\" (UniqueName: \"kubernetes.io/projected/b9146130-53f7-4c79-adff-c093da73c967-kube-api-access-bqmd8\") on node \"crc\" DevicePath \"\"" Dec 06 03:53:22 crc kubenswrapper[4980]: I1206 03:53:22.428997 4980 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/b9146130-53f7-4c79-adff-c093da73c967-etc-swift\") on node \"crc\" DevicePath \"\"" Dec 06 03:53:22 crc kubenswrapper[4980]: I1206 03:53:22.429014 4980 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b9146130-53f7-4c79-adff-c093da73c967-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 03:53:22 crc kubenswrapper[4980]: I1206 03:53:22.541831 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/swift-ring-rebalance-pfjgb" Dec 06 03:53:23 crc kubenswrapper[4980]: I1206 03:53:23.036797 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/92027099-00cc-45db-bbfe-7796fe5a68c4-etc-swift\") pod \"swift-proxy-85fcf7dd9-sn4m2\" (UID: \"92027099-00cc-45db-bbfe-7796fe5a68c4\") " pod="glance-kuttl-tests/swift-proxy-85fcf7dd9-sn4m2" Dec 06 03:53:23 crc kubenswrapper[4980]: E1206 03:53:23.037010 4980 projected.go:288] Couldn't get configMap glance-kuttl-tests/swift-ring-files: configmap "swift-ring-files" not found Dec 06 03:53:23 crc kubenswrapper[4980]: E1206 03:53:23.037204 4980 projected.go:194] Error preparing data for projected volume etc-swift for pod glance-kuttl-tests/swift-proxy-85fcf7dd9-sn4m2: configmap "swift-ring-files" not found Dec 06 03:53:23 crc kubenswrapper[4980]: E1206 03:53:23.037277 4980 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/92027099-00cc-45db-bbfe-7796fe5a68c4-etc-swift podName:92027099-00cc-45db-bbfe-7796fe5a68c4 nodeName:}" failed. No retries permitted until 2025-12-06 03:53:31.037255654 +0000 UTC m=+1170.282006925 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/92027099-00cc-45db-bbfe-7796fe5a68c4-etc-swift") pod "swift-proxy-85fcf7dd9-sn4m2" (UID: "92027099-00cc-45db-bbfe-7796fe5a68c4") : configmap "swift-ring-files" not found Dec 06 03:53:23 crc kubenswrapper[4980]: I1206 03:53:23.037828 4980 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/swift-ring-rebalance-pfjgb"] Dec 06 03:53:23 crc kubenswrapper[4980]: W1206 03:53:23.047502 4980 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod43cddc9e_ad31_4596_8e89_cb3260fd55a8.slice/crio-07a504e78033b279d06381d55657462d506d3b9432d833aa16e480a89d9d4252 WatchSource:0}: Error finding container 07a504e78033b279d06381d55657462d506d3b9432d833aa16e480a89d9d4252: Status 404 returned error can't find the container with id 07a504e78033b279d06381d55657462d506d3b9432d833aa16e480a89d9d4252 Dec 06 03:53:23 crc kubenswrapper[4980]: I1206 03:53:23.140052 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-ring-rebalance-pfjgb" event={"ID":"43cddc9e-ad31-4596-8e89-cb3260fd55a8","Type":"ContainerStarted","Data":"07a504e78033b279d06381d55657462d506d3b9432d833aa16e480a89d9d4252"} Dec 06 03:53:23 crc kubenswrapper[4980]: I1206 03:53:23.140085 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/swift-ring-rebalance-vn7bt" Dec 06 03:53:23 crc kubenswrapper[4980]: I1206 03:53:23.207308 4980 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/swift-ring-rebalance-vn7bt"] Dec 06 03:53:23 crc kubenswrapper[4980]: I1206 03:53:23.222791 4980 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/swift-ring-rebalance-vn7bt"] Dec 06 03:53:25 crc kubenswrapper[4980]: I1206 03:53:25.137499 4980 patch_prober.go:28] interesting pod/machine-config-daemon-r5zfb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 03:53:25 crc kubenswrapper[4980]: I1206 03:53:25.137994 4980 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-r5zfb" podUID="320f44d1-a671-4a91-b328-a8b0fdd8f23a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 03:53:25 crc kubenswrapper[4980]: I1206 03:53:25.138056 4980 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-r5zfb" Dec 06 03:53:25 crc kubenswrapper[4980]: I1206 03:53:25.138812 4980 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"a9db17110188a0ff97c6851d4bf74c82d9f9edeef9b7cfcb0fd15528f7d2e059"} pod="openshift-machine-config-operator/machine-config-daemon-r5zfb" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 06 03:53:25 crc kubenswrapper[4980]: I1206 03:53:25.138885 4980 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-r5zfb" podUID="320f44d1-a671-4a91-b328-a8b0fdd8f23a" containerName="machine-config-daemon" containerID="cri-o://a9db17110188a0ff97c6851d4bf74c82d9f9edeef9b7cfcb0fd15528f7d2e059" gracePeriod=600 Dec 06 03:53:25 crc kubenswrapper[4980]: I1206 03:53:25.167687 4980 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/glance-operator-index-bzpqg" Dec 06 03:53:25 crc kubenswrapper[4980]: I1206 03:53:25.167746 4980 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/glance-operator-index-bzpqg" Dec 06 03:53:25 crc kubenswrapper[4980]: I1206 03:53:25.197273 4980 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b9146130-53f7-4c79-adff-c093da73c967" path="/var/lib/kubelet/pods/b9146130-53f7-4c79-adff-c093da73c967/volumes" Dec 06 03:53:25 crc kubenswrapper[4980]: I1206 03:53:25.200965 4980 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/glance-operator-index-bzpqg" Dec 06 03:53:26 crc kubenswrapper[4980]: I1206 03:53:26.172474 4980 generic.go:334] "Generic (PLEG): container finished" podID="320f44d1-a671-4a91-b328-a8b0fdd8f23a" containerID="a9db17110188a0ff97c6851d4bf74c82d9f9edeef9b7cfcb0fd15528f7d2e059" exitCode=0 Dec 06 03:53:26 crc kubenswrapper[4980]: I1206 03:53:26.172554 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-r5zfb" event={"ID":"320f44d1-a671-4a91-b328-a8b0fdd8f23a","Type":"ContainerDied","Data":"a9db17110188a0ff97c6851d4bf74c82d9f9edeef9b7cfcb0fd15528f7d2e059"} Dec 06 03:53:26 crc kubenswrapper[4980]: I1206 03:53:26.173002 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-r5zfb" event={"ID":"320f44d1-a671-4a91-b328-a8b0fdd8f23a","Type":"ContainerStarted","Data":"ed3956e825d000e94b97025c20ede41bab44553dd92465ad73c08a9ec44729eb"} Dec 06 03:53:26 crc kubenswrapper[4980]: I1206 03:53:26.173031 4980 scope.go:117] "RemoveContainer" containerID="89e5bc79489f00d58aeba62f0124b8200c624808c64c4cf50ac8bd11cba63d00" Dec 06 03:53:26 crc kubenswrapper[4980]: I1206 03:53:26.220705 4980 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/glance-operator-index-bzpqg" Dec 06 03:53:26 crc kubenswrapper[4980]: I1206 03:53:26.300628 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/3752191f-1260-4be2-8d68-308491d20468-etc-swift\") pod \"swift-storage-0\" (UID: \"3752191f-1260-4be2-8d68-308491d20468\") " pod="glance-kuttl-tests/swift-storage-0" Dec 06 03:53:26 crc kubenswrapper[4980]: E1206 03:53:26.300782 4980 projected.go:288] Couldn't get configMap glance-kuttl-tests/swift-ring-files: configmap "swift-ring-files" not found Dec 06 03:53:26 crc kubenswrapper[4980]: E1206 03:53:26.300803 4980 projected.go:194] Error preparing data for projected volume etc-swift for pod glance-kuttl-tests/swift-storage-0: configmap "swift-ring-files" not found Dec 06 03:53:26 crc kubenswrapper[4980]: E1206 03:53:26.300861 4980 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3752191f-1260-4be2-8d68-308491d20468-etc-swift podName:3752191f-1260-4be2-8d68-308491d20468 nodeName:}" failed. No retries permitted until 2025-12-06 03:53:34.300846851 +0000 UTC m=+1173.545598122 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/3752191f-1260-4be2-8d68-308491d20468-etc-swift") pod "swift-storage-0" (UID: "3752191f-1260-4be2-8d68-308491d20468") : configmap "swift-ring-files" not found Dec 06 03:53:29 crc kubenswrapper[4980]: I1206 03:53:29.200580 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-ring-rebalance-pfjgb" event={"ID":"43cddc9e-ad31-4596-8e89-cb3260fd55a8","Type":"ContainerStarted","Data":"52b2f4da5595b5712d31482fd58946e9bd724f30dc462a859841457685e3ff58"} Dec 06 03:53:29 crc kubenswrapper[4980]: I1206 03:53:29.221264 4980 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/swift-ring-rebalance-pfjgb" podStartSLOduration=2.9372792199999997 podStartE2EDuration="8.221199519s" podCreationTimestamp="2025-12-06 03:53:21 +0000 UTC" firstStartedPulling="2025-12-06 03:53:23.050214657 +0000 UTC m=+1162.294965918" lastFinishedPulling="2025-12-06 03:53:28.334134946 +0000 UTC m=+1167.578886217" observedRunningTime="2025-12-06 03:53:29.215411652 +0000 UTC m=+1168.460162963" watchObservedRunningTime="2025-12-06 03:53:29.221199519 +0000 UTC m=+1168.465950830" Dec 06 03:53:31 crc kubenswrapper[4980]: I1206 03:53:31.118410 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/92027099-00cc-45db-bbfe-7796fe5a68c4-etc-swift\") pod \"swift-proxy-85fcf7dd9-sn4m2\" (UID: \"92027099-00cc-45db-bbfe-7796fe5a68c4\") " pod="glance-kuttl-tests/swift-proxy-85fcf7dd9-sn4m2" Dec 06 03:53:31 crc kubenswrapper[4980]: E1206 03:53:31.118587 4980 projected.go:288] Couldn't get configMap glance-kuttl-tests/swift-ring-files: configmap "swift-ring-files" not found Dec 06 03:53:31 crc kubenswrapper[4980]: E1206 03:53:31.118983 4980 projected.go:194] Error preparing data for projected volume etc-swift for pod glance-kuttl-tests/swift-proxy-85fcf7dd9-sn4m2: configmap "swift-ring-files" not found Dec 06 03:53:31 crc kubenswrapper[4980]: E1206 03:53:31.119044 4980 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/92027099-00cc-45db-bbfe-7796fe5a68c4-etc-swift podName:92027099-00cc-45db-bbfe-7796fe5a68c4 nodeName:}" failed. No retries permitted until 2025-12-06 03:53:47.119027267 +0000 UTC m=+1186.363778538 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/92027099-00cc-45db-bbfe-7796fe5a68c4-etc-swift") pod "swift-proxy-85fcf7dd9-sn4m2" (UID: "92027099-00cc-45db-bbfe-7796fe5a68c4") : configmap "swift-ring-files" not found Dec 06 03:53:34 crc kubenswrapper[4980]: I1206 03:53:34.364277 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/3752191f-1260-4be2-8d68-308491d20468-etc-swift\") pod \"swift-storage-0\" (UID: \"3752191f-1260-4be2-8d68-308491d20468\") " pod="glance-kuttl-tests/swift-storage-0" Dec 06 03:53:34 crc kubenswrapper[4980]: E1206 03:53:34.364582 4980 projected.go:288] Couldn't get configMap glance-kuttl-tests/swift-ring-files: configmap "swift-ring-files" not found Dec 06 03:53:34 crc kubenswrapper[4980]: E1206 03:53:34.365063 4980 projected.go:194] Error preparing data for projected volume etc-swift for pod glance-kuttl-tests/swift-storage-0: configmap "swift-ring-files" not found Dec 06 03:53:34 crc kubenswrapper[4980]: E1206 03:53:34.365141 4980 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3752191f-1260-4be2-8d68-308491d20468-etc-swift podName:3752191f-1260-4be2-8d68-308491d20468 nodeName:}" failed. No retries permitted until 2025-12-06 03:53:50.365112775 +0000 UTC m=+1189.609864066 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/3752191f-1260-4be2-8d68-308491d20468-etc-swift") pod "swift-storage-0" (UID: "3752191f-1260-4be2-8d68-308491d20468") : configmap "swift-ring-files" not found Dec 06 03:53:35 crc kubenswrapper[4980]: I1206 03:53:35.299222 4980 generic.go:334] "Generic (PLEG): container finished" podID="43cddc9e-ad31-4596-8e89-cb3260fd55a8" containerID="52b2f4da5595b5712d31482fd58946e9bd724f30dc462a859841457685e3ff58" exitCode=0 Dec 06 03:53:35 crc kubenswrapper[4980]: I1206 03:53:35.299297 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-ring-rebalance-pfjgb" event={"ID":"43cddc9e-ad31-4596-8e89-cb3260fd55a8","Type":"ContainerDied","Data":"52b2f4da5595b5712d31482fd58946e9bd724f30dc462a859841457685e3ff58"} Dec 06 03:53:36 crc kubenswrapper[4980]: I1206 03:53:36.642765 4980 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/swift-ring-rebalance-pfjgb" Dec 06 03:53:36 crc kubenswrapper[4980]: I1206 03:53:36.818122 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/43cddc9e-ad31-4596-8e89-cb3260fd55a8-swiftconf\") pod \"43cddc9e-ad31-4596-8e89-cb3260fd55a8\" (UID: \"43cddc9e-ad31-4596-8e89-cb3260fd55a8\") " Dec 06 03:53:36 crc kubenswrapper[4980]: I1206 03:53:36.818211 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/43cddc9e-ad31-4596-8e89-cb3260fd55a8-etc-swift\") pod \"43cddc9e-ad31-4596-8e89-cb3260fd55a8\" (UID: \"43cddc9e-ad31-4596-8e89-cb3260fd55a8\") " Dec 06 03:53:36 crc kubenswrapper[4980]: I1206 03:53:36.818323 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/43cddc9e-ad31-4596-8e89-cb3260fd55a8-ring-data-devices\") pod \"43cddc9e-ad31-4596-8e89-cb3260fd55a8\" (UID: \"43cddc9e-ad31-4596-8e89-cb3260fd55a8\") " Dec 06 03:53:36 crc kubenswrapper[4980]: I1206 03:53:36.818353 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/43cddc9e-ad31-4596-8e89-cb3260fd55a8-scripts\") pod \"43cddc9e-ad31-4596-8e89-cb3260fd55a8\" (UID: \"43cddc9e-ad31-4596-8e89-cb3260fd55a8\") " Dec 06 03:53:36 crc kubenswrapper[4980]: I1206 03:53:36.818370 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hdpzd\" (UniqueName: \"kubernetes.io/projected/43cddc9e-ad31-4596-8e89-cb3260fd55a8-kube-api-access-hdpzd\") pod \"43cddc9e-ad31-4596-8e89-cb3260fd55a8\" (UID: \"43cddc9e-ad31-4596-8e89-cb3260fd55a8\") " Dec 06 03:53:36 crc kubenswrapper[4980]: I1206 03:53:36.818388 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/43cddc9e-ad31-4596-8e89-cb3260fd55a8-dispersionconf\") pod \"43cddc9e-ad31-4596-8e89-cb3260fd55a8\" (UID: \"43cddc9e-ad31-4596-8e89-cb3260fd55a8\") " Dec 06 03:53:36 crc kubenswrapper[4980]: I1206 03:53:36.819752 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43cddc9e-ad31-4596-8e89-cb3260fd55a8-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "43cddc9e-ad31-4596-8e89-cb3260fd55a8" (UID: "43cddc9e-ad31-4596-8e89-cb3260fd55a8"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 03:53:36 crc kubenswrapper[4980]: I1206 03:53:36.820365 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/43cddc9e-ad31-4596-8e89-cb3260fd55a8-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "43cddc9e-ad31-4596-8e89-cb3260fd55a8" (UID: "43cddc9e-ad31-4596-8e89-cb3260fd55a8"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 03:53:36 crc kubenswrapper[4980]: I1206 03:53:36.823865 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/43cddc9e-ad31-4596-8e89-cb3260fd55a8-kube-api-access-hdpzd" (OuterVolumeSpecName: "kube-api-access-hdpzd") pod "43cddc9e-ad31-4596-8e89-cb3260fd55a8" (UID: "43cddc9e-ad31-4596-8e89-cb3260fd55a8"). InnerVolumeSpecName "kube-api-access-hdpzd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 03:53:36 crc kubenswrapper[4980]: I1206 03:53:36.837535 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43cddc9e-ad31-4596-8e89-cb3260fd55a8-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "43cddc9e-ad31-4596-8e89-cb3260fd55a8" (UID: "43cddc9e-ad31-4596-8e89-cb3260fd55a8"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 03:53:36 crc kubenswrapper[4980]: I1206 03:53:36.839022 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43cddc9e-ad31-4596-8e89-cb3260fd55a8-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "43cddc9e-ad31-4596-8e89-cb3260fd55a8" (UID: "43cddc9e-ad31-4596-8e89-cb3260fd55a8"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 03:53:36 crc kubenswrapper[4980]: I1206 03:53:36.842157 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43cddc9e-ad31-4596-8e89-cb3260fd55a8-scripts" (OuterVolumeSpecName: "scripts") pod "43cddc9e-ad31-4596-8e89-cb3260fd55a8" (UID: "43cddc9e-ad31-4596-8e89-cb3260fd55a8"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 03:53:36 crc kubenswrapper[4980]: I1206 03:53:36.920144 4980 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/43cddc9e-ad31-4596-8e89-cb3260fd55a8-etc-swift\") on node \"crc\" DevicePath \"\"" Dec 06 03:53:36 crc kubenswrapper[4980]: I1206 03:53:36.920177 4980 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/43cddc9e-ad31-4596-8e89-cb3260fd55a8-ring-data-devices\") on node \"crc\" DevicePath \"\"" Dec 06 03:53:36 crc kubenswrapper[4980]: I1206 03:53:36.920192 4980 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/43cddc9e-ad31-4596-8e89-cb3260fd55a8-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 03:53:36 crc kubenswrapper[4980]: I1206 03:53:36.920207 4980 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hdpzd\" (UniqueName: \"kubernetes.io/projected/43cddc9e-ad31-4596-8e89-cb3260fd55a8-kube-api-access-hdpzd\") on node \"crc\" DevicePath \"\"" Dec 06 03:53:36 crc kubenswrapper[4980]: I1206 03:53:36.920219 4980 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/43cddc9e-ad31-4596-8e89-cb3260fd55a8-dispersionconf\") on node \"crc\" DevicePath \"\"" Dec 06 03:53:36 crc kubenswrapper[4980]: I1206 03:53:36.920231 4980 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/43cddc9e-ad31-4596-8e89-cb3260fd55a8-swiftconf\") on node \"crc\" DevicePath \"\"" Dec 06 03:53:37 crc kubenswrapper[4980]: I1206 03:53:37.312863 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-ring-rebalance-pfjgb" event={"ID":"43cddc9e-ad31-4596-8e89-cb3260fd55a8","Type":"ContainerDied","Data":"07a504e78033b279d06381d55657462d506d3b9432d833aa16e480a89d9d4252"} Dec 06 03:53:37 crc kubenswrapper[4980]: I1206 03:53:37.312905 4980 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="07a504e78033b279d06381d55657462d506d3b9432d833aa16e480a89d9d4252" Dec 06 03:53:37 crc kubenswrapper[4980]: I1206 03:53:37.312926 4980 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/swift-ring-rebalance-pfjgb" Dec 06 03:53:44 crc kubenswrapper[4980]: I1206 03:53:44.690889 4980 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/82cf678a38d30dcf5cc31c16a6d10360b45b3df2f5eaffd238882752a25cbw6"] Dec 06 03:53:44 crc kubenswrapper[4980]: E1206 03:53:44.692697 4980 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="43cddc9e-ad31-4596-8e89-cb3260fd55a8" containerName="swift-ring-rebalance" Dec 06 03:53:44 crc kubenswrapper[4980]: I1206 03:53:44.692713 4980 state_mem.go:107] "Deleted CPUSet assignment" podUID="43cddc9e-ad31-4596-8e89-cb3260fd55a8" containerName="swift-ring-rebalance" Dec 06 03:53:44 crc kubenswrapper[4980]: I1206 03:53:44.692848 4980 memory_manager.go:354] "RemoveStaleState removing state" podUID="43cddc9e-ad31-4596-8e89-cb3260fd55a8" containerName="swift-ring-rebalance" Dec 06 03:53:44 crc kubenswrapper[4980]: I1206 03:53:44.694196 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/82cf678a38d30dcf5cc31c16a6d10360b45b3df2f5eaffd238882752a25cbw6" Dec 06 03:53:44 crc kubenswrapper[4980]: I1206 03:53:44.697679 4980 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"default-dockercfg-brs8h" Dec 06 03:53:44 crc kubenswrapper[4980]: I1206 03:53:44.719931 4980 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/82cf678a38d30dcf5cc31c16a6d10360b45b3df2f5eaffd238882752a25cbw6"] Dec 06 03:53:44 crc kubenswrapper[4980]: I1206 03:53:44.839794 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tpjzn\" (UniqueName: \"kubernetes.io/projected/37352474-24e2-4b44-b5d7-c482f8cffa3f-kube-api-access-tpjzn\") pod \"82cf678a38d30dcf5cc31c16a6d10360b45b3df2f5eaffd238882752a25cbw6\" (UID: \"37352474-24e2-4b44-b5d7-c482f8cffa3f\") " pod="openstack-operators/82cf678a38d30dcf5cc31c16a6d10360b45b3df2f5eaffd238882752a25cbw6" Dec 06 03:53:44 crc kubenswrapper[4980]: I1206 03:53:44.840167 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/37352474-24e2-4b44-b5d7-c482f8cffa3f-bundle\") pod \"82cf678a38d30dcf5cc31c16a6d10360b45b3df2f5eaffd238882752a25cbw6\" (UID: \"37352474-24e2-4b44-b5d7-c482f8cffa3f\") " pod="openstack-operators/82cf678a38d30dcf5cc31c16a6d10360b45b3df2f5eaffd238882752a25cbw6" Dec 06 03:53:44 crc kubenswrapper[4980]: I1206 03:53:44.840195 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/37352474-24e2-4b44-b5d7-c482f8cffa3f-util\") pod \"82cf678a38d30dcf5cc31c16a6d10360b45b3df2f5eaffd238882752a25cbw6\" (UID: \"37352474-24e2-4b44-b5d7-c482f8cffa3f\") " pod="openstack-operators/82cf678a38d30dcf5cc31c16a6d10360b45b3df2f5eaffd238882752a25cbw6" Dec 06 03:53:44 crc kubenswrapper[4980]: I1206 03:53:44.941122 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tpjzn\" (UniqueName: \"kubernetes.io/projected/37352474-24e2-4b44-b5d7-c482f8cffa3f-kube-api-access-tpjzn\") pod \"82cf678a38d30dcf5cc31c16a6d10360b45b3df2f5eaffd238882752a25cbw6\" (UID: \"37352474-24e2-4b44-b5d7-c482f8cffa3f\") " pod="openstack-operators/82cf678a38d30dcf5cc31c16a6d10360b45b3df2f5eaffd238882752a25cbw6" Dec 06 03:53:44 crc kubenswrapper[4980]: I1206 03:53:44.941185 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/37352474-24e2-4b44-b5d7-c482f8cffa3f-bundle\") pod \"82cf678a38d30dcf5cc31c16a6d10360b45b3df2f5eaffd238882752a25cbw6\" (UID: \"37352474-24e2-4b44-b5d7-c482f8cffa3f\") " pod="openstack-operators/82cf678a38d30dcf5cc31c16a6d10360b45b3df2f5eaffd238882752a25cbw6" Dec 06 03:53:44 crc kubenswrapper[4980]: I1206 03:53:44.941214 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/37352474-24e2-4b44-b5d7-c482f8cffa3f-util\") pod \"82cf678a38d30dcf5cc31c16a6d10360b45b3df2f5eaffd238882752a25cbw6\" (UID: \"37352474-24e2-4b44-b5d7-c482f8cffa3f\") " pod="openstack-operators/82cf678a38d30dcf5cc31c16a6d10360b45b3df2f5eaffd238882752a25cbw6" Dec 06 03:53:44 crc kubenswrapper[4980]: I1206 03:53:44.941822 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/37352474-24e2-4b44-b5d7-c482f8cffa3f-util\") pod \"82cf678a38d30dcf5cc31c16a6d10360b45b3df2f5eaffd238882752a25cbw6\" (UID: \"37352474-24e2-4b44-b5d7-c482f8cffa3f\") " pod="openstack-operators/82cf678a38d30dcf5cc31c16a6d10360b45b3df2f5eaffd238882752a25cbw6" Dec 06 03:53:44 crc kubenswrapper[4980]: I1206 03:53:44.941872 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/37352474-24e2-4b44-b5d7-c482f8cffa3f-bundle\") pod \"82cf678a38d30dcf5cc31c16a6d10360b45b3df2f5eaffd238882752a25cbw6\" (UID: \"37352474-24e2-4b44-b5d7-c482f8cffa3f\") " pod="openstack-operators/82cf678a38d30dcf5cc31c16a6d10360b45b3df2f5eaffd238882752a25cbw6" Dec 06 03:53:44 crc kubenswrapper[4980]: I1206 03:53:44.958394 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tpjzn\" (UniqueName: \"kubernetes.io/projected/37352474-24e2-4b44-b5d7-c482f8cffa3f-kube-api-access-tpjzn\") pod \"82cf678a38d30dcf5cc31c16a6d10360b45b3df2f5eaffd238882752a25cbw6\" (UID: \"37352474-24e2-4b44-b5d7-c482f8cffa3f\") " pod="openstack-operators/82cf678a38d30dcf5cc31c16a6d10360b45b3df2f5eaffd238882752a25cbw6" Dec 06 03:53:45 crc kubenswrapper[4980]: I1206 03:53:45.022371 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/82cf678a38d30dcf5cc31c16a6d10360b45b3df2f5eaffd238882752a25cbw6" Dec 06 03:53:45 crc kubenswrapper[4980]: I1206 03:53:45.454571 4980 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/82cf678a38d30dcf5cc31c16a6d10360b45b3df2f5eaffd238882752a25cbw6"] Dec 06 03:53:46 crc kubenswrapper[4980]: I1206 03:53:46.372274 4980 generic.go:334] "Generic (PLEG): container finished" podID="37352474-24e2-4b44-b5d7-c482f8cffa3f" containerID="ea26ec2f655271c06cc2e73ec91f5e64690b4fa0c211c5b5039d7bddb9c38cff" exitCode=0 Dec 06 03:53:46 crc kubenswrapper[4980]: I1206 03:53:46.372362 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/82cf678a38d30dcf5cc31c16a6d10360b45b3df2f5eaffd238882752a25cbw6" event={"ID":"37352474-24e2-4b44-b5d7-c482f8cffa3f","Type":"ContainerDied","Data":"ea26ec2f655271c06cc2e73ec91f5e64690b4fa0c211c5b5039d7bddb9c38cff"} Dec 06 03:53:46 crc kubenswrapper[4980]: I1206 03:53:46.372638 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/82cf678a38d30dcf5cc31c16a6d10360b45b3df2f5eaffd238882752a25cbw6" event={"ID":"37352474-24e2-4b44-b5d7-c482f8cffa3f","Type":"ContainerStarted","Data":"f375e09c80c8a7a8c802e9064e5e7ad0d85f53d81626cb7ee3c82d0be3a4975d"} Dec 06 03:53:47 crc kubenswrapper[4980]: I1206 03:53:47.182324 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/92027099-00cc-45db-bbfe-7796fe5a68c4-etc-swift\") pod \"swift-proxy-85fcf7dd9-sn4m2\" (UID: \"92027099-00cc-45db-bbfe-7796fe5a68c4\") " pod="glance-kuttl-tests/swift-proxy-85fcf7dd9-sn4m2" Dec 06 03:53:47 crc kubenswrapper[4980]: I1206 03:53:47.191319 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/92027099-00cc-45db-bbfe-7796fe5a68c4-etc-swift\") pod \"swift-proxy-85fcf7dd9-sn4m2\" (UID: \"92027099-00cc-45db-bbfe-7796fe5a68c4\") " pod="glance-kuttl-tests/swift-proxy-85fcf7dd9-sn4m2" Dec 06 03:53:47 crc kubenswrapper[4980]: I1206 03:53:47.319598 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/swift-proxy-85fcf7dd9-sn4m2" Dec 06 03:53:47 crc kubenswrapper[4980]: I1206 03:53:47.739335 4980 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/swift-proxy-85fcf7dd9-sn4m2"] Dec 06 03:53:47 crc kubenswrapper[4980]: W1206 03:53:47.829132 4980 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod92027099_00cc_45db_bbfe_7796fe5a68c4.slice/crio-f8478da8dd9ef7993612d5bd285f9a85e824fc205e9a243eacbcb162d4ce006f WatchSource:0}: Error finding container f8478da8dd9ef7993612d5bd285f9a85e824fc205e9a243eacbcb162d4ce006f: Status 404 returned error can't find the container with id f8478da8dd9ef7993612d5bd285f9a85e824fc205e9a243eacbcb162d4ce006f Dec 06 03:53:48 crc kubenswrapper[4980]: I1206 03:53:48.395333 4980 generic.go:334] "Generic (PLEG): container finished" podID="37352474-24e2-4b44-b5d7-c482f8cffa3f" containerID="fd438197b2b898d6cdb0261f1c8432e10c742fb2691f18bf27f7e463d2ee8870" exitCode=0 Dec 06 03:53:48 crc kubenswrapper[4980]: I1206 03:53:48.395412 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/82cf678a38d30dcf5cc31c16a6d10360b45b3df2f5eaffd238882752a25cbw6" event={"ID":"37352474-24e2-4b44-b5d7-c482f8cffa3f","Type":"ContainerDied","Data":"fd438197b2b898d6cdb0261f1c8432e10c742fb2691f18bf27f7e463d2ee8870"} Dec 06 03:53:48 crc kubenswrapper[4980]: I1206 03:53:48.399712 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-proxy-85fcf7dd9-sn4m2" event={"ID":"92027099-00cc-45db-bbfe-7796fe5a68c4","Type":"ContainerStarted","Data":"e84a0b7af9cf80245662ce92b81046b9a83aa1478db47fec6aee3a882e935ec2"} Dec 06 03:53:48 crc kubenswrapper[4980]: I1206 03:53:48.399758 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-proxy-85fcf7dd9-sn4m2" event={"ID":"92027099-00cc-45db-bbfe-7796fe5a68c4","Type":"ContainerStarted","Data":"81a126d630e98daea7a610d66a6d0b1a82e5954a2d36c923ba87ed4724677639"} Dec 06 03:53:48 crc kubenswrapper[4980]: I1206 03:53:48.399771 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-proxy-85fcf7dd9-sn4m2" event={"ID":"92027099-00cc-45db-bbfe-7796fe5a68c4","Type":"ContainerStarted","Data":"f8478da8dd9ef7993612d5bd285f9a85e824fc205e9a243eacbcb162d4ce006f"} Dec 06 03:53:48 crc kubenswrapper[4980]: I1206 03:53:48.400226 4980 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/swift-proxy-85fcf7dd9-sn4m2" Dec 06 03:53:48 crc kubenswrapper[4980]: I1206 03:53:48.400272 4980 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/swift-proxy-85fcf7dd9-sn4m2" Dec 06 03:53:48 crc kubenswrapper[4980]: I1206 03:53:48.452048 4980 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/swift-proxy-85fcf7dd9-sn4m2" podStartSLOduration=33.452020769 podStartE2EDuration="33.452020769s" podCreationTimestamp="2025-12-06 03:53:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 03:53:48.449300581 +0000 UTC m=+1187.694051852" watchObservedRunningTime="2025-12-06 03:53:48.452020769 +0000 UTC m=+1187.696772060" Dec 06 03:53:49 crc kubenswrapper[4980]: I1206 03:53:49.408592 4980 generic.go:334] "Generic (PLEG): container finished" podID="37352474-24e2-4b44-b5d7-c482f8cffa3f" containerID="c62e8c89c5f423784437a6ae52d9dba9ed4ae0813e3d7f3b7adf1fbb4b1ae0a5" exitCode=0 Dec 06 03:53:49 crc kubenswrapper[4980]: I1206 03:53:49.409661 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/82cf678a38d30dcf5cc31c16a6d10360b45b3df2f5eaffd238882752a25cbw6" event={"ID":"37352474-24e2-4b44-b5d7-c482f8cffa3f","Type":"ContainerDied","Data":"c62e8c89c5f423784437a6ae52d9dba9ed4ae0813e3d7f3b7adf1fbb4b1ae0a5"} Dec 06 03:53:50 crc kubenswrapper[4980]: I1206 03:53:50.433066 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/3752191f-1260-4be2-8d68-308491d20468-etc-swift\") pod \"swift-storage-0\" (UID: \"3752191f-1260-4be2-8d68-308491d20468\") " pod="glance-kuttl-tests/swift-storage-0" Dec 06 03:53:50 crc kubenswrapper[4980]: I1206 03:53:50.445007 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/3752191f-1260-4be2-8d68-308491d20468-etc-swift\") pod \"swift-storage-0\" (UID: \"3752191f-1260-4be2-8d68-308491d20468\") " pod="glance-kuttl-tests/swift-storage-0" Dec 06 03:53:50 crc kubenswrapper[4980]: I1206 03:53:50.502286 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/swift-storage-0" Dec 06 03:53:50 crc kubenswrapper[4980]: I1206 03:53:50.701029 4980 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/82cf678a38d30dcf5cc31c16a6d10360b45b3df2f5eaffd238882752a25cbw6" Dec 06 03:53:50 crc kubenswrapper[4980]: I1206 03:53:50.840320 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/37352474-24e2-4b44-b5d7-c482f8cffa3f-bundle\") pod \"37352474-24e2-4b44-b5d7-c482f8cffa3f\" (UID: \"37352474-24e2-4b44-b5d7-c482f8cffa3f\") " Dec 06 03:53:50 crc kubenswrapper[4980]: I1206 03:53:50.840461 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/37352474-24e2-4b44-b5d7-c482f8cffa3f-util\") pod \"37352474-24e2-4b44-b5d7-c482f8cffa3f\" (UID: \"37352474-24e2-4b44-b5d7-c482f8cffa3f\") " Dec 06 03:53:50 crc kubenswrapper[4980]: I1206 03:53:50.840574 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tpjzn\" (UniqueName: \"kubernetes.io/projected/37352474-24e2-4b44-b5d7-c482f8cffa3f-kube-api-access-tpjzn\") pod \"37352474-24e2-4b44-b5d7-c482f8cffa3f\" (UID: \"37352474-24e2-4b44-b5d7-c482f8cffa3f\") " Dec 06 03:53:50 crc kubenswrapper[4980]: I1206 03:53:50.841400 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/37352474-24e2-4b44-b5d7-c482f8cffa3f-bundle" (OuterVolumeSpecName: "bundle") pod "37352474-24e2-4b44-b5d7-c482f8cffa3f" (UID: "37352474-24e2-4b44-b5d7-c482f8cffa3f"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 03:53:50 crc kubenswrapper[4980]: I1206 03:53:50.845862 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/37352474-24e2-4b44-b5d7-c482f8cffa3f-kube-api-access-tpjzn" (OuterVolumeSpecName: "kube-api-access-tpjzn") pod "37352474-24e2-4b44-b5d7-c482f8cffa3f" (UID: "37352474-24e2-4b44-b5d7-c482f8cffa3f"). InnerVolumeSpecName "kube-api-access-tpjzn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 03:53:50 crc kubenswrapper[4980]: I1206 03:53:50.855247 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/37352474-24e2-4b44-b5d7-c482f8cffa3f-util" (OuterVolumeSpecName: "util") pod "37352474-24e2-4b44-b5d7-c482f8cffa3f" (UID: "37352474-24e2-4b44-b5d7-c482f8cffa3f"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 03:53:50 crc kubenswrapper[4980]: I1206 03:53:50.942191 4980 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/37352474-24e2-4b44-b5d7-c482f8cffa3f-util\") on node \"crc\" DevicePath \"\"" Dec 06 03:53:50 crc kubenswrapper[4980]: I1206 03:53:50.942269 4980 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tpjzn\" (UniqueName: \"kubernetes.io/projected/37352474-24e2-4b44-b5d7-c482f8cffa3f-kube-api-access-tpjzn\") on node \"crc\" DevicePath \"\"" Dec 06 03:53:50 crc kubenswrapper[4980]: I1206 03:53:50.942299 4980 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/37352474-24e2-4b44-b5d7-c482f8cffa3f-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 03:53:50 crc kubenswrapper[4980]: I1206 03:53:50.992479 4980 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/swift-storage-0"] Dec 06 03:53:51 crc kubenswrapper[4980]: W1206 03:53:51.001842 4980 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3752191f_1260_4be2_8d68_308491d20468.slice/crio-6b11c316703281b98c03fab21ab474593f42b425c50d1dec6b48a9d7a68d2218 WatchSource:0}: Error finding container 6b11c316703281b98c03fab21ab474593f42b425c50d1dec6b48a9d7a68d2218: Status 404 returned error can't find the container with id 6b11c316703281b98c03fab21ab474593f42b425c50d1dec6b48a9d7a68d2218 Dec 06 03:53:51 crc kubenswrapper[4980]: I1206 03:53:51.423506 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-storage-0" event={"ID":"3752191f-1260-4be2-8d68-308491d20468","Type":"ContainerStarted","Data":"6b11c316703281b98c03fab21ab474593f42b425c50d1dec6b48a9d7a68d2218"} Dec 06 03:53:51 crc kubenswrapper[4980]: I1206 03:53:51.427043 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/82cf678a38d30dcf5cc31c16a6d10360b45b3df2f5eaffd238882752a25cbw6" event={"ID":"37352474-24e2-4b44-b5d7-c482f8cffa3f","Type":"ContainerDied","Data":"f375e09c80c8a7a8c802e9064e5e7ad0d85f53d81626cb7ee3c82d0be3a4975d"} Dec 06 03:53:51 crc kubenswrapper[4980]: I1206 03:53:51.427097 4980 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f375e09c80c8a7a8c802e9064e5e7ad0d85f53d81626cb7ee3c82d0be3a4975d" Dec 06 03:53:51 crc kubenswrapper[4980]: I1206 03:53:51.427110 4980 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/82cf678a38d30dcf5cc31c16a6d10360b45b3df2f5eaffd238882752a25cbw6" Dec 06 03:53:54 crc kubenswrapper[4980]: I1206 03:53:54.457859 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-storage-0" event={"ID":"3752191f-1260-4be2-8d68-308491d20468","Type":"ContainerStarted","Data":"4f7e7a9954c718aa5eab68719430f74597abaec8078317d64781e4c3ef76804b"} Dec 06 03:53:54 crc kubenswrapper[4980]: I1206 03:53:54.458192 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-storage-0" event={"ID":"3752191f-1260-4be2-8d68-308491d20468","Type":"ContainerStarted","Data":"5e5cb017aaf7f1a926c8da29e22d604aa62c8d39936afab0f6cb6797da7e9077"} Dec 06 03:53:54 crc kubenswrapper[4980]: I1206 03:53:54.458219 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-storage-0" event={"ID":"3752191f-1260-4be2-8d68-308491d20468","Type":"ContainerStarted","Data":"ab0b99eed260a530094cd399fd88fbe237500926190ac2d1ed8bbf7fc35f7496"} Dec 06 03:53:54 crc kubenswrapper[4980]: I1206 03:53:54.458235 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-storage-0" event={"ID":"3752191f-1260-4be2-8d68-308491d20468","Type":"ContainerStarted","Data":"872f2ce9c2227a9650b01ec9bc723cd5188f99887c7644979253ad7cd8f21e9a"} Dec 06 03:53:57 crc kubenswrapper[4980]: I1206 03:53:57.323882 4980 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/swift-proxy-85fcf7dd9-sn4m2" Dec 06 03:53:57 crc kubenswrapper[4980]: I1206 03:53:57.325900 4980 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/swift-proxy-85fcf7dd9-sn4m2" Dec 06 03:53:57 crc kubenswrapper[4980]: I1206 03:53:57.483495 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-storage-0" event={"ID":"3752191f-1260-4be2-8d68-308491d20468","Type":"ContainerStarted","Data":"367bbf40ed7893ce0b9e7f411ac21ae1b99b4aa4b022a767c5b4d71302fec8dd"} Dec 06 03:53:57 crc kubenswrapper[4980]: I1206 03:53:57.483774 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-storage-0" event={"ID":"3752191f-1260-4be2-8d68-308491d20468","Type":"ContainerStarted","Data":"b4885e60fe85f7bb0d51034c327695c614e342218e65f34d4b42be1a3264fee6"} Dec 06 03:53:57 crc kubenswrapper[4980]: I1206 03:53:57.483784 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-storage-0" event={"ID":"3752191f-1260-4be2-8d68-308491d20468","Type":"ContainerStarted","Data":"4bd1af62fdcee29e725e561737f031d8ec936adcf064d98291e586b59bf0422c"} Dec 06 03:53:57 crc kubenswrapper[4980]: I1206 03:53:57.483792 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-storage-0" event={"ID":"3752191f-1260-4be2-8d68-308491d20468","Type":"ContainerStarted","Data":"a2d73c2b25ac1dd51edb0a702c78e98dfadca47b50226a78a9caef156b78f83a"} Dec 06 03:53:59 crc kubenswrapper[4980]: I1206 03:53:59.548087 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-storage-0" event={"ID":"3752191f-1260-4be2-8d68-308491d20468","Type":"ContainerStarted","Data":"2f6623695190a2e4436a59bd2700336390a2e7622d50c34d23802232fb02f28d"} Dec 06 03:54:00 crc kubenswrapper[4980]: I1206 03:54:00.566374 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-storage-0" event={"ID":"3752191f-1260-4be2-8d68-308491d20468","Type":"ContainerStarted","Data":"d5c597fa563cb254b471a1f0c095d37eff39a2fa66edf6986cb8b9e922bdb188"} Dec 06 03:54:00 crc kubenswrapper[4980]: I1206 03:54:00.566722 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-storage-0" event={"ID":"3752191f-1260-4be2-8d68-308491d20468","Type":"ContainerStarted","Data":"8af0cc0ac7995e072d0ad18196a8b22e342ab66fd7b01b77e819b5bb1acf11dc"} Dec 06 03:54:01 crc kubenswrapper[4980]: I1206 03:54:01.580452 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-storage-0" event={"ID":"3752191f-1260-4be2-8d68-308491d20468","Type":"ContainerStarted","Data":"191dd83fa136a347e7f7a6a26c596b7cc365160d34ca924ef55eeee0f10611c5"} Dec 06 03:54:01 crc kubenswrapper[4980]: I1206 03:54:01.580747 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-storage-0" event={"ID":"3752191f-1260-4be2-8d68-308491d20468","Type":"ContainerStarted","Data":"dcaba756d378e38093ba71b1bc4515df98a5476c05ddda4de50ed31be072d64f"} Dec 06 03:54:01 crc kubenswrapper[4980]: I1206 03:54:01.580758 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-storage-0" event={"ID":"3752191f-1260-4be2-8d68-308491d20468","Type":"ContainerStarted","Data":"b8da31ad5ae4621dcbc20417165ec7c27a6c4f76bfcb3c00e7f270ccb29dde5a"} Dec 06 03:54:01 crc kubenswrapper[4980]: I1206 03:54:01.580766 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-storage-0" event={"ID":"3752191f-1260-4be2-8d68-308491d20468","Type":"ContainerStarted","Data":"eeceb4a9ddb6694adf6e7ccecc5f8309d59b4b86bfc1aacf256e64e3fd83b995"} Dec 06 03:54:01 crc kubenswrapper[4980]: I1206 03:54:01.621723 4980 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/swift-storage-0" podStartSLOduration=36.47637811 podStartE2EDuration="44.621687684s" podCreationTimestamp="2025-12-06 03:53:17 +0000 UTC" firstStartedPulling="2025-12-06 03:53:51.00441247 +0000 UTC m=+1190.249163741" lastFinishedPulling="2025-12-06 03:53:59.149722044 +0000 UTC m=+1198.394473315" observedRunningTime="2025-12-06 03:54:01.612674635 +0000 UTC m=+1200.857425906" watchObservedRunningTime="2025-12-06 03:54:01.621687684 +0000 UTC m=+1200.866438955" Dec 06 03:54:12 crc kubenswrapper[4980]: I1206 03:54:12.541583 4980 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/glance-operator-controller-manager-549766cbfc-nxztk"] Dec 06 03:54:12 crc kubenswrapper[4980]: E1206 03:54:12.543305 4980 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="37352474-24e2-4b44-b5d7-c482f8cffa3f" containerName="util" Dec 06 03:54:12 crc kubenswrapper[4980]: I1206 03:54:12.543398 4980 state_mem.go:107] "Deleted CPUSet assignment" podUID="37352474-24e2-4b44-b5d7-c482f8cffa3f" containerName="util" Dec 06 03:54:12 crc kubenswrapper[4980]: E1206 03:54:12.543477 4980 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="37352474-24e2-4b44-b5d7-c482f8cffa3f" containerName="pull" Dec 06 03:54:12 crc kubenswrapper[4980]: I1206 03:54:12.543563 4980 state_mem.go:107] "Deleted CPUSet assignment" podUID="37352474-24e2-4b44-b5d7-c482f8cffa3f" containerName="pull" Dec 06 03:54:12 crc kubenswrapper[4980]: E1206 03:54:12.543642 4980 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="37352474-24e2-4b44-b5d7-c482f8cffa3f" containerName="extract" Dec 06 03:54:12 crc kubenswrapper[4980]: I1206 03:54:12.543706 4980 state_mem.go:107] "Deleted CPUSet assignment" podUID="37352474-24e2-4b44-b5d7-c482f8cffa3f" containerName="extract" Dec 06 03:54:12 crc kubenswrapper[4980]: I1206 03:54:12.543905 4980 memory_manager.go:354] "RemoveStaleState removing state" podUID="37352474-24e2-4b44-b5d7-c482f8cffa3f" containerName="extract" Dec 06 03:54:12 crc kubenswrapper[4980]: I1206 03:54:12.544872 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-549766cbfc-nxztk" Dec 06 03:54:12 crc kubenswrapper[4980]: I1206 03:54:12.548098 4980 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"glance-operator-controller-manager-service-cert" Dec 06 03:54:12 crc kubenswrapper[4980]: I1206 03:54:12.565951 4980 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"glance-operator-controller-manager-dockercfg-wjxv4" Dec 06 03:54:12 crc kubenswrapper[4980]: I1206 03:54:12.587152 4980 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-549766cbfc-nxztk"] Dec 06 03:54:12 crc kubenswrapper[4980]: I1206 03:54:12.672996 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/c28fde7c-003e-45ed-bb75-fec17db60f64-webhook-cert\") pod \"glance-operator-controller-manager-549766cbfc-nxztk\" (UID: \"c28fde7c-003e-45ed-bb75-fec17db60f64\") " pod="openstack-operators/glance-operator-controller-manager-549766cbfc-nxztk" Dec 06 03:54:12 crc kubenswrapper[4980]: I1206 03:54:12.673114 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/c28fde7c-003e-45ed-bb75-fec17db60f64-apiservice-cert\") pod \"glance-operator-controller-manager-549766cbfc-nxztk\" (UID: \"c28fde7c-003e-45ed-bb75-fec17db60f64\") " pod="openstack-operators/glance-operator-controller-manager-549766cbfc-nxztk" Dec 06 03:54:12 crc kubenswrapper[4980]: I1206 03:54:12.673171 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-98wlq\" (UniqueName: \"kubernetes.io/projected/c28fde7c-003e-45ed-bb75-fec17db60f64-kube-api-access-98wlq\") pod \"glance-operator-controller-manager-549766cbfc-nxztk\" (UID: \"c28fde7c-003e-45ed-bb75-fec17db60f64\") " pod="openstack-operators/glance-operator-controller-manager-549766cbfc-nxztk" Dec 06 03:54:12 crc kubenswrapper[4980]: I1206 03:54:12.773711 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/c28fde7c-003e-45ed-bb75-fec17db60f64-webhook-cert\") pod \"glance-operator-controller-manager-549766cbfc-nxztk\" (UID: \"c28fde7c-003e-45ed-bb75-fec17db60f64\") " pod="openstack-operators/glance-operator-controller-manager-549766cbfc-nxztk" Dec 06 03:54:12 crc kubenswrapper[4980]: I1206 03:54:12.774628 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/c28fde7c-003e-45ed-bb75-fec17db60f64-apiservice-cert\") pod \"glance-operator-controller-manager-549766cbfc-nxztk\" (UID: \"c28fde7c-003e-45ed-bb75-fec17db60f64\") " pod="openstack-operators/glance-operator-controller-manager-549766cbfc-nxztk" Dec 06 03:54:12 crc kubenswrapper[4980]: I1206 03:54:12.774698 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-98wlq\" (UniqueName: \"kubernetes.io/projected/c28fde7c-003e-45ed-bb75-fec17db60f64-kube-api-access-98wlq\") pod \"glance-operator-controller-manager-549766cbfc-nxztk\" (UID: \"c28fde7c-003e-45ed-bb75-fec17db60f64\") " pod="openstack-operators/glance-operator-controller-manager-549766cbfc-nxztk" Dec 06 03:54:12 crc kubenswrapper[4980]: I1206 03:54:12.780056 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/c28fde7c-003e-45ed-bb75-fec17db60f64-apiservice-cert\") pod \"glance-operator-controller-manager-549766cbfc-nxztk\" (UID: \"c28fde7c-003e-45ed-bb75-fec17db60f64\") " pod="openstack-operators/glance-operator-controller-manager-549766cbfc-nxztk" Dec 06 03:54:12 crc kubenswrapper[4980]: I1206 03:54:12.780717 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/c28fde7c-003e-45ed-bb75-fec17db60f64-webhook-cert\") pod \"glance-operator-controller-manager-549766cbfc-nxztk\" (UID: \"c28fde7c-003e-45ed-bb75-fec17db60f64\") " pod="openstack-operators/glance-operator-controller-manager-549766cbfc-nxztk" Dec 06 03:54:12 crc kubenswrapper[4980]: I1206 03:54:12.792093 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-98wlq\" (UniqueName: \"kubernetes.io/projected/c28fde7c-003e-45ed-bb75-fec17db60f64-kube-api-access-98wlq\") pod \"glance-operator-controller-manager-549766cbfc-nxztk\" (UID: \"c28fde7c-003e-45ed-bb75-fec17db60f64\") " pod="openstack-operators/glance-operator-controller-manager-549766cbfc-nxztk" Dec 06 03:54:12 crc kubenswrapper[4980]: I1206 03:54:12.888342 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-549766cbfc-nxztk" Dec 06 03:54:13 crc kubenswrapper[4980]: I1206 03:54:13.149443 4980 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-549766cbfc-nxztk"] Dec 06 03:54:13 crc kubenswrapper[4980]: W1206 03:54:13.154288 4980 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc28fde7c_003e_45ed_bb75_fec17db60f64.slice/crio-7c32bf1dd59709778bb23877aad3eecf848e54b310dc5804b7636a669e955c95 WatchSource:0}: Error finding container 7c32bf1dd59709778bb23877aad3eecf848e54b310dc5804b7636a669e955c95: Status 404 returned error can't find the container with id 7c32bf1dd59709778bb23877aad3eecf848e54b310dc5804b7636a669e955c95 Dec 06 03:54:13 crc kubenswrapper[4980]: I1206 03:54:13.683252 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-549766cbfc-nxztk" event={"ID":"c28fde7c-003e-45ed-bb75-fec17db60f64","Type":"ContainerStarted","Data":"7c32bf1dd59709778bb23877aad3eecf848e54b310dc5804b7636a669e955c95"} Dec 06 03:54:16 crc kubenswrapper[4980]: I1206 03:54:16.761419 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-549766cbfc-nxztk" event={"ID":"c28fde7c-003e-45ed-bb75-fec17db60f64","Type":"ContainerStarted","Data":"ef86d0f78876e3b1109fd47ce01de5437e470971abf3b19b6381c879b403c63c"} Dec 06 03:54:16 crc kubenswrapper[4980]: I1206 03:54:16.762057 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-549766cbfc-nxztk" event={"ID":"c28fde7c-003e-45ed-bb75-fec17db60f64","Type":"ContainerStarted","Data":"812d304da669dd1e7358a190f7e28be2651d8e1561a570642269f169c3588be4"} Dec 06 03:54:16 crc kubenswrapper[4980]: I1206 03:54:16.762644 4980 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/glance-operator-controller-manager-549766cbfc-nxztk" Dec 06 03:54:16 crc kubenswrapper[4980]: I1206 03:54:16.798654 4980 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/glance-operator-controller-manager-549766cbfc-nxztk" podStartSLOduration=1.771987077 podStartE2EDuration="4.798636412s" podCreationTimestamp="2025-12-06 03:54:12 +0000 UTC" firstStartedPulling="2025-12-06 03:54:13.15695841 +0000 UTC m=+1212.401709681" lastFinishedPulling="2025-12-06 03:54:16.183607725 +0000 UTC m=+1215.428359016" observedRunningTime="2025-12-06 03:54:16.793962658 +0000 UTC m=+1216.038713939" watchObservedRunningTime="2025-12-06 03:54:16.798636412 +0000 UTC m=+1216.043387683" Dec 06 03:54:22 crc kubenswrapper[4980]: I1206 03:54:22.895377 4980 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/glance-operator-controller-manager-549766cbfc-nxztk" Dec 06 03:54:28 crc kubenswrapper[4980]: I1206 03:54:28.483593 4980 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-db-create-whc99"] Dec 06 03:54:28 crc kubenswrapper[4980]: I1206 03:54:28.485559 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-create-whc99" Dec 06 03:54:28 crc kubenswrapper[4980]: I1206 03:54:28.497950 4980 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-db-create-whc99"] Dec 06 03:54:28 crc kubenswrapper[4980]: I1206 03:54:28.515675 4980 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/openstackclient"] Dec 06 03:54:28 crc kubenswrapper[4980]: I1206 03:54:28.516384 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/openstackclient" Dec 06 03:54:28 crc kubenswrapper[4980]: I1206 03:54:28.523886 4980 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/openstackclient"] Dec 06 03:54:28 crc kubenswrapper[4980]: I1206 03:54:28.524246 4980 reflector.go:368] Caches populated for *v1.ConfigMap from object-"glance-kuttl-tests"/"openstack-config" Dec 06 03:54:28 crc kubenswrapper[4980]: I1206 03:54:28.524462 4980 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"default-dockercfg-mc277" Dec 06 03:54:28 crc kubenswrapper[4980]: I1206 03:54:28.524595 4980 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"openstack-config-secret" Dec 06 03:54:28 crc kubenswrapper[4980]: I1206 03:54:28.524808 4980 reflector.go:368] Caches populated for *v1.ConfigMap from object-"glance-kuttl-tests"/"openstack-scripts-9db6gc427h" Dec 06 03:54:28 crc kubenswrapper[4980]: I1206 03:54:28.528068 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jsmrb\" (UniqueName: \"kubernetes.io/projected/50b592be-1d9d-45c7-9cc9-4c7082d2af78-kube-api-access-jsmrb\") pod \"glance-db-create-whc99\" (UID: \"50b592be-1d9d-45c7-9cc9-4c7082d2af78\") " pod="glance-kuttl-tests/glance-db-create-whc99" Dec 06 03:54:28 crc kubenswrapper[4980]: I1206 03:54:28.630091 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-scripts\" (UniqueName: \"kubernetes.io/configmap/7c853f1b-b0d8-4f38-9ad3-0dcb3b2e0707-openstack-scripts\") pod \"openstackclient\" (UID: \"7c853f1b-b0d8-4f38-9ad3-0dcb3b2e0707\") " pod="glance-kuttl-tests/openstackclient" Dec 06 03:54:28 crc kubenswrapper[4980]: I1206 03:54:28.630572 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/7c853f1b-b0d8-4f38-9ad3-0dcb3b2e0707-openstack-config-secret\") pod \"openstackclient\" (UID: \"7c853f1b-b0d8-4f38-9ad3-0dcb3b2e0707\") " pod="glance-kuttl-tests/openstackclient" Dec 06 03:54:28 crc kubenswrapper[4980]: I1206 03:54:28.630766 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/7c853f1b-b0d8-4f38-9ad3-0dcb3b2e0707-openstack-config\") pod \"openstackclient\" (UID: \"7c853f1b-b0d8-4f38-9ad3-0dcb3b2e0707\") " pod="glance-kuttl-tests/openstackclient" Dec 06 03:54:28 crc kubenswrapper[4980]: I1206 03:54:28.630935 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5rrbt\" (UniqueName: \"kubernetes.io/projected/7c853f1b-b0d8-4f38-9ad3-0dcb3b2e0707-kube-api-access-5rrbt\") pod \"openstackclient\" (UID: \"7c853f1b-b0d8-4f38-9ad3-0dcb3b2e0707\") " pod="glance-kuttl-tests/openstackclient" Dec 06 03:54:28 crc kubenswrapper[4980]: I1206 03:54:28.631141 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jsmrb\" (UniqueName: \"kubernetes.io/projected/50b592be-1d9d-45c7-9cc9-4c7082d2af78-kube-api-access-jsmrb\") pod \"glance-db-create-whc99\" (UID: \"50b592be-1d9d-45c7-9cc9-4c7082d2af78\") " pod="glance-kuttl-tests/glance-db-create-whc99" Dec 06 03:54:28 crc kubenswrapper[4980]: I1206 03:54:28.651288 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jsmrb\" (UniqueName: \"kubernetes.io/projected/50b592be-1d9d-45c7-9cc9-4c7082d2af78-kube-api-access-jsmrb\") pod \"glance-db-create-whc99\" (UID: \"50b592be-1d9d-45c7-9cc9-4c7082d2af78\") " pod="glance-kuttl-tests/glance-db-create-whc99" Dec 06 03:54:28 crc kubenswrapper[4980]: I1206 03:54:28.733067 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/7c853f1b-b0d8-4f38-9ad3-0dcb3b2e0707-openstack-config\") pod \"openstackclient\" (UID: \"7c853f1b-b0d8-4f38-9ad3-0dcb3b2e0707\") " pod="glance-kuttl-tests/openstackclient" Dec 06 03:54:28 crc kubenswrapper[4980]: I1206 03:54:28.733127 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5rrbt\" (UniqueName: \"kubernetes.io/projected/7c853f1b-b0d8-4f38-9ad3-0dcb3b2e0707-kube-api-access-5rrbt\") pod \"openstackclient\" (UID: \"7c853f1b-b0d8-4f38-9ad3-0dcb3b2e0707\") " pod="glance-kuttl-tests/openstackclient" Dec 06 03:54:28 crc kubenswrapper[4980]: I1206 03:54:28.733188 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-scripts\" (UniqueName: \"kubernetes.io/configmap/7c853f1b-b0d8-4f38-9ad3-0dcb3b2e0707-openstack-scripts\") pod \"openstackclient\" (UID: \"7c853f1b-b0d8-4f38-9ad3-0dcb3b2e0707\") " pod="glance-kuttl-tests/openstackclient" Dec 06 03:54:28 crc kubenswrapper[4980]: I1206 03:54:28.733262 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/7c853f1b-b0d8-4f38-9ad3-0dcb3b2e0707-openstack-config-secret\") pod \"openstackclient\" (UID: \"7c853f1b-b0d8-4f38-9ad3-0dcb3b2e0707\") " pod="glance-kuttl-tests/openstackclient" Dec 06 03:54:28 crc kubenswrapper[4980]: I1206 03:54:28.734974 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/7c853f1b-b0d8-4f38-9ad3-0dcb3b2e0707-openstack-config\") pod \"openstackclient\" (UID: \"7c853f1b-b0d8-4f38-9ad3-0dcb3b2e0707\") " pod="glance-kuttl-tests/openstackclient" Dec 06 03:54:28 crc kubenswrapper[4980]: I1206 03:54:28.735612 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-scripts\" (UniqueName: \"kubernetes.io/configmap/7c853f1b-b0d8-4f38-9ad3-0dcb3b2e0707-openstack-scripts\") pod \"openstackclient\" (UID: \"7c853f1b-b0d8-4f38-9ad3-0dcb3b2e0707\") " pod="glance-kuttl-tests/openstackclient" Dec 06 03:54:28 crc kubenswrapper[4980]: I1206 03:54:28.738695 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/7c853f1b-b0d8-4f38-9ad3-0dcb3b2e0707-openstack-config-secret\") pod \"openstackclient\" (UID: \"7c853f1b-b0d8-4f38-9ad3-0dcb3b2e0707\") " pod="glance-kuttl-tests/openstackclient" Dec 06 03:54:28 crc kubenswrapper[4980]: I1206 03:54:28.754534 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5rrbt\" (UniqueName: \"kubernetes.io/projected/7c853f1b-b0d8-4f38-9ad3-0dcb3b2e0707-kube-api-access-5rrbt\") pod \"openstackclient\" (UID: \"7c853f1b-b0d8-4f38-9ad3-0dcb3b2e0707\") " pod="glance-kuttl-tests/openstackclient" Dec 06 03:54:28 crc kubenswrapper[4980]: I1206 03:54:28.845370 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-create-whc99" Dec 06 03:54:28 crc kubenswrapper[4980]: I1206 03:54:28.860385 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/openstackclient" Dec 06 03:54:29 crc kubenswrapper[4980]: I1206 03:54:29.296255 4980 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-db-create-whc99"] Dec 06 03:54:29 crc kubenswrapper[4980]: I1206 03:54:29.333983 4980 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/openstackclient"] Dec 06 03:54:29 crc kubenswrapper[4980]: W1206 03:54:29.358182 4980 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7c853f1b_b0d8_4f38_9ad3_0dcb3b2e0707.slice/crio-ef6423d15175d50287cb4ccb582811f976e235b548f1c363508a945cd94d83c7 WatchSource:0}: Error finding container ef6423d15175d50287cb4ccb582811f976e235b548f1c363508a945cd94d83c7: Status 404 returned error can't find the container with id ef6423d15175d50287cb4ccb582811f976e235b548f1c363508a945cd94d83c7 Dec 06 03:54:29 crc kubenswrapper[4980]: I1206 03:54:29.883065 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/openstackclient" event={"ID":"7c853f1b-b0d8-4f38-9ad3-0dcb3b2e0707","Type":"ContainerStarted","Data":"ef6423d15175d50287cb4ccb582811f976e235b548f1c363508a945cd94d83c7"} Dec 06 03:54:29 crc kubenswrapper[4980]: I1206 03:54:29.886113 4980 generic.go:334] "Generic (PLEG): container finished" podID="50b592be-1d9d-45c7-9cc9-4c7082d2af78" containerID="e372798becf1c32343fcbf6a7cf618cccedec38ee0dce588b1a6f1db47fc3d6e" exitCode=0 Dec 06 03:54:29 crc kubenswrapper[4980]: I1206 03:54:29.886269 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-create-whc99" event={"ID":"50b592be-1d9d-45c7-9cc9-4c7082d2af78","Type":"ContainerDied","Data":"e372798becf1c32343fcbf6a7cf618cccedec38ee0dce588b1a6f1db47fc3d6e"} Dec 06 03:54:29 crc kubenswrapper[4980]: I1206 03:54:29.886388 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-create-whc99" event={"ID":"50b592be-1d9d-45c7-9cc9-4c7082d2af78","Type":"ContainerStarted","Data":"523bba8efa00edf64cd14fb7acb0616c376e0748db01807dfdeab2df25628ba1"} Dec 06 03:54:31 crc kubenswrapper[4980]: I1206 03:54:31.223597 4980 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-create-whc99" Dec 06 03:54:31 crc kubenswrapper[4980]: I1206 03:54:31.271722 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jsmrb\" (UniqueName: \"kubernetes.io/projected/50b592be-1d9d-45c7-9cc9-4c7082d2af78-kube-api-access-jsmrb\") pod \"50b592be-1d9d-45c7-9cc9-4c7082d2af78\" (UID: \"50b592be-1d9d-45c7-9cc9-4c7082d2af78\") " Dec 06 03:54:31 crc kubenswrapper[4980]: I1206 03:54:31.276981 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/50b592be-1d9d-45c7-9cc9-4c7082d2af78-kube-api-access-jsmrb" (OuterVolumeSpecName: "kube-api-access-jsmrb") pod "50b592be-1d9d-45c7-9cc9-4c7082d2af78" (UID: "50b592be-1d9d-45c7-9cc9-4c7082d2af78"). InnerVolumeSpecName "kube-api-access-jsmrb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 03:54:31 crc kubenswrapper[4980]: I1206 03:54:31.373487 4980 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jsmrb\" (UniqueName: \"kubernetes.io/projected/50b592be-1d9d-45c7-9cc9-4c7082d2af78-kube-api-access-jsmrb\") on node \"crc\" DevicePath \"\"" Dec 06 03:54:31 crc kubenswrapper[4980]: I1206 03:54:31.902697 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-create-whc99" event={"ID":"50b592be-1d9d-45c7-9cc9-4c7082d2af78","Type":"ContainerDied","Data":"523bba8efa00edf64cd14fb7acb0616c376e0748db01807dfdeab2df25628ba1"} Dec 06 03:54:31 crc kubenswrapper[4980]: I1206 03:54:31.903037 4980 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="523bba8efa00edf64cd14fb7acb0616c376e0748db01807dfdeab2df25628ba1" Dec 06 03:54:31 crc kubenswrapper[4980]: I1206 03:54:31.902890 4980 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-create-whc99" Dec 06 03:54:38 crc kubenswrapper[4980]: I1206 03:54:38.499248 4980 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-f836-account-create-9h4q6"] Dec 06 03:54:38 crc kubenswrapper[4980]: E1206 03:54:38.500551 4980 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="50b592be-1d9d-45c7-9cc9-4c7082d2af78" containerName="mariadb-database-create" Dec 06 03:54:38 crc kubenswrapper[4980]: I1206 03:54:38.500572 4980 state_mem.go:107] "Deleted CPUSet assignment" podUID="50b592be-1d9d-45c7-9cc9-4c7082d2af78" containerName="mariadb-database-create" Dec 06 03:54:38 crc kubenswrapper[4980]: I1206 03:54:38.501635 4980 memory_manager.go:354] "RemoveStaleState removing state" podUID="50b592be-1d9d-45c7-9cc9-4c7082d2af78" containerName="mariadb-database-create" Dec 06 03:54:38 crc kubenswrapper[4980]: I1206 03:54:38.502306 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-f836-account-create-9h4q6" Dec 06 03:54:38 crc kubenswrapper[4980]: I1206 03:54:38.504347 4980 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-db-secret" Dec 06 03:54:38 crc kubenswrapper[4980]: I1206 03:54:38.505790 4980 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-f836-account-create-9h4q6"] Dec 06 03:54:38 crc kubenswrapper[4980]: I1206 03:54:38.639500 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8prlc\" (UniqueName: \"kubernetes.io/projected/7575d3d1-f9b7-4840-bca9-30d6eaa5d0c3-kube-api-access-8prlc\") pod \"glance-f836-account-create-9h4q6\" (UID: \"7575d3d1-f9b7-4840-bca9-30d6eaa5d0c3\") " pod="glance-kuttl-tests/glance-f836-account-create-9h4q6" Dec 06 03:54:38 crc kubenswrapper[4980]: I1206 03:54:38.741210 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8prlc\" (UniqueName: \"kubernetes.io/projected/7575d3d1-f9b7-4840-bca9-30d6eaa5d0c3-kube-api-access-8prlc\") pod \"glance-f836-account-create-9h4q6\" (UID: \"7575d3d1-f9b7-4840-bca9-30d6eaa5d0c3\") " pod="glance-kuttl-tests/glance-f836-account-create-9h4q6" Dec 06 03:54:38 crc kubenswrapper[4980]: I1206 03:54:38.759193 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8prlc\" (UniqueName: \"kubernetes.io/projected/7575d3d1-f9b7-4840-bca9-30d6eaa5d0c3-kube-api-access-8prlc\") pod \"glance-f836-account-create-9h4q6\" (UID: \"7575d3d1-f9b7-4840-bca9-30d6eaa5d0c3\") " pod="glance-kuttl-tests/glance-f836-account-create-9h4q6" Dec 06 03:54:38 crc kubenswrapper[4980]: I1206 03:54:38.824840 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-f836-account-create-9h4q6" Dec 06 03:54:39 crc kubenswrapper[4980]: I1206 03:54:39.359011 4980 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-f836-account-create-9h4q6"] Dec 06 03:54:39 crc kubenswrapper[4980]: W1206 03:54:39.359351 4980 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7575d3d1_f9b7_4840_bca9_30d6eaa5d0c3.slice/crio-780dd81881f9bb121f9eb5e1fa36869cd0af33987328ffbc0e7eecf5f377eb4b WatchSource:0}: Error finding container 780dd81881f9bb121f9eb5e1fa36869cd0af33987328ffbc0e7eecf5f377eb4b: Status 404 returned error can't find the container with id 780dd81881f9bb121f9eb5e1fa36869cd0af33987328ffbc0e7eecf5f377eb4b Dec 06 03:54:39 crc kubenswrapper[4980]: I1206 03:54:39.978076 4980 generic.go:334] "Generic (PLEG): container finished" podID="7575d3d1-f9b7-4840-bca9-30d6eaa5d0c3" containerID="4e6f6016172688199f5ee8bb2fbe6e702d804a3bf782f709b474dbfe051a3c94" exitCode=0 Dec 06 03:54:39 crc kubenswrapper[4980]: I1206 03:54:39.978143 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-f836-account-create-9h4q6" event={"ID":"7575d3d1-f9b7-4840-bca9-30d6eaa5d0c3","Type":"ContainerDied","Data":"4e6f6016172688199f5ee8bb2fbe6e702d804a3bf782f709b474dbfe051a3c94"} Dec 06 03:54:39 crc kubenswrapper[4980]: I1206 03:54:39.978380 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-f836-account-create-9h4q6" event={"ID":"7575d3d1-f9b7-4840-bca9-30d6eaa5d0c3","Type":"ContainerStarted","Data":"780dd81881f9bb121f9eb5e1fa36869cd0af33987328ffbc0e7eecf5f377eb4b"} Dec 06 03:54:39 crc kubenswrapper[4980]: I1206 03:54:39.982542 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/openstackclient" event={"ID":"7c853f1b-b0d8-4f38-9ad3-0dcb3b2e0707","Type":"ContainerStarted","Data":"5a6c6907ac4004409028d10f1c5c0615205912ae630e496046598525b59ccad4"} Dec 06 03:54:40 crc kubenswrapper[4980]: I1206 03:54:40.011157 4980 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/openstackclient" podStartSLOduration=2.372476608 podStartE2EDuration="12.011130951s" podCreationTimestamp="2025-12-06 03:54:28 +0000 UTC" firstStartedPulling="2025-12-06 03:54:29.36093409 +0000 UTC m=+1228.605685361" lastFinishedPulling="2025-12-06 03:54:38.999588433 +0000 UTC m=+1238.244339704" observedRunningTime="2025-12-06 03:54:40.007133286 +0000 UTC m=+1239.251884557" watchObservedRunningTime="2025-12-06 03:54:40.011130951 +0000 UTC m=+1239.255882222" Dec 06 03:54:41 crc kubenswrapper[4980]: I1206 03:54:41.287010 4980 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-f836-account-create-9h4q6" Dec 06 03:54:41 crc kubenswrapper[4980]: I1206 03:54:41.382995 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8prlc\" (UniqueName: \"kubernetes.io/projected/7575d3d1-f9b7-4840-bca9-30d6eaa5d0c3-kube-api-access-8prlc\") pod \"7575d3d1-f9b7-4840-bca9-30d6eaa5d0c3\" (UID: \"7575d3d1-f9b7-4840-bca9-30d6eaa5d0c3\") " Dec 06 03:54:41 crc kubenswrapper[4980]: I1206 03:54:41.388011 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7575d3d1-f9b7-4840-bca9-30d6eaa5d0c3-kube-api-access-8prlc" (OuterVolumeSpecName: "kube-api-access-8prlc") pod "7575d3d1-f9b7-4840-bca9-30d6eaa5d0c3" (UID: "7575d3d1-f9b7-4840-bca9-30d6eaa5d0c3"). InnerVolumeSpecName "kube-api-access-8prlc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 03:54:41 crc kubenswrapper[4980]: I1206 03:54:41.484379 4980 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8prlc\" (UniqueName: \"kubernetes.io/projected/7575d3d1-f9b7-4840-bca9-30d6eaa5d0c3-kube-api-access-8prlc\") on node \"crc\" DevicePath \"\"" Dec 06 03:54:41 crc kubenswrapper[4980]: I1206 03:54:41.998825 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-f836-account-create-9h4q6" event={"ID":"7575d3d1-f9b7-4840-bca9-30d6eaa5d0c3","Type":"ContainerDied","Data":"780dd81881f9bb121f9eb5e1fa36869cd0af33987328ffbc0e7eecf5f377eb4b"} Dec 06 03:54:41 crc kubenswrapper[4980]: I1206 03:54:41.998899 4980 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="780dd81881f9bb121f9eb5e1fa36869cd0af33987328ffbc0e7eecf5f377eb4b" Dec 06 03:54:41 crc kubenswrapper[4980]: I1206 03:54:41.998979 4980 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-f836-account-create-9h4q6" Dec 06 03:54:43 crc kubenswrapper[4980]: I1206 03:54:43.627556 4980 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-db-sync-bdqlq"] Dec 06 03:54:43 crc kubenswrapper[4980]: E1206 03:54:43.628158 4980 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7575d3d1-f9b7-4840-bca9-30d6eaa5d0c3" containerName="mariadb-account-create" Dec 06 03:54:43 crc kubenswrapper[4980]: I1206 03:54:43.628175 4980 state_mem.go:107] "Deleted CPUSet assignment" podUID="7575d3d1-f9b7-4840-bca9-30d6eaa5d0c3" containerName="mariadb-account-create" Dec 06 03:54:43 crc kubenswrapper[4980]: I1206 03:54:43.628371 4980 memory_manager.go:354] "RemoveStaleState removing state" podUID="7575d3d1-f9b7-4840-bca9-30d6eaa5d0c3" containerName="mariadb-account-create" Dec 06 03:54:43 crc kubenswrapper[4980]: I1206 03:54:43.628986 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-sync-bdqlq" Dec 06 03:54:43 crc kubenswrapper[4980]: I1206 03:54:43.631468 4980 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-config-data" Dec 06 03:54:43 crc kubenswrapper[4980]: I1206 03:54:43.632167 4980 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-glance-dockercfg-kpxrp" Dec 06 03:54:43 crc kubenswrapper[4980]: I1206 03:54:43.643627 4980 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-db-sync-bdqlq"] Dec 06 03:54:43 crc kubenswrapper[4980]: I1206 03:54:43.714476 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d8db6f70-28bc-4b58-91e9-feb5bd06a3e4-config-data\") pod \"glance-db-sync-bdqlq\" (UID: \"d8db6f70-28bc-4b58-91e9-feb5bd06a3e4\") " pod="glance-kuttl-tests/glance-db-sync-bdqlq" Dec 06 03:54:43 crc kubenswrapper[4980]: I1206 03:54:43.714541 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5rh6z\" (UniqueName: \"kubernetes.io/projected/d8db6f70-28bc-4b58-91e9-feb5bd06a3e4-kube-api-access-5rh6z\") pod \"glance-db-sync-bdqlq\" (UID: \"d8db6f70-28bc-4b58-91e9-feb5bd06a3e4\") " pod="glance-kuttl-tests/glance-db-sync-bdqlq" Dec 06 03:54:43 crc kubenswrapper[4980]: I1206 03:54:43.714581 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/d8db6f70-28bc-4b58-91e9-feb5bd06a3e4-db-sync-config-data\") pod \"glance-db-sync-bdqlq\" (UID: \"d8db6f70-28bc-4b58-91e9-feb5bd06a3e4\") " pod="glance-kuttl-tests/glance-db-sync-bdqlq" Dec 06 03:54:43 crc kubenswrapper[4980]: I1206 03:54:43.815966 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d8db6f70-28bc-4b58-91e9-feb5bd06a3e4-config-data\") pod \"glance-db-sync-bdqlq\" (UID: \"d8db6f70-28bc-4b58-91e9-feb5bd06a3e4\") " pod="glance-kuttl-tests/glance-db-sync-bdqlq" Dec 06 03:54:43 crc kubenswrapper[4980]: I1206 03:54:43.816312 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5rh6z\" (UniqueName: \"kubernetes.io/projected/d8db6f70-28bc-4b58-91e9-feb5bd06a3e4-kube-api-access-5rh6z\") pod \"glance-db-sync-bdqlq\" (UID: \"d8db6f70-28bc-4b58-91e9-feb5bd06a3e4\") " pod="glance-kuttl-tests/glance-db-sync-bdqlq" Dec 06 03:54:43 crc kubenswrapper[4980]: I1206 03:54:43.816473 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/d8db6f70-28bc-4b58-91e9-feb5bd06a3e4-db-sync-config-data\") pod \"glance-db-sync-bdqlq\" (UID: \"d8db6f70-28bc-4b58-91e9-feb5bd06a3e4\") " pod="glance-kuttl-tests/glance-db-sync-bdqlq" Dec 06 03:54:43 crc kubenswrapper[4980]: I1206 03:54:43.821652 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d8db6f70-28bc-4b58-91e9-feb5bd06a3e4-config-data\") pod \"glance-db-sync-bdqlq\" (UID: \"d8db6f70-28bc-4b58-91e9-feb5bd06a3e4\") " pod="glance-kuttl-tests/glance-db-sync-bdqlq" Dec 06 03:54:43 crc kubenswrapper[4980]: I1206 03:54:43.826748 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/d8db6f70-28bc-4b58-91e9-feb5bd06a3e4-db-sync-config-data\") pod \"glance-db-sync-bdqlq\" (UID: \"d8db6f70-28bc-4b58-91e9-feb5bd06a3e4\") " pod="glance-kuttl-tests/glance-db-sync-bdqlq" Dec 06 03:54:43 crc kubenswrapper[4980]: I1206 03:54:43.843292 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5rh6z\" (UniqueName: \"kubernetes.io/projected/d8db6f70-28bc-4b58-91e9-feb5bd06a3e4-kube-api-access-5rh6z\") pod \"glance-db-sync-bdqlq\" (UID: \"d8db6f70-28bc-4b58-91e9-feb5bd06a3e4\") " pod="glance-kuttl-tests/glance-db-sync-bdqlq" Dec 06 03:54:43 crc kubenswrapper[4980]: I1206 03:54:43.958060 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-sync-bdqlq" Dec 06 03:54:44 crc kubenswrapper[4980]: I1206 03:54:44.191882 4980 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-db-sync-bdqlq"] Dec 06 03:54:44 crc kubenswrapper[4980]: W1206 03:54:44.200772 4980 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd8db6f70_28bc_4b58_91e9_feb5bd06a3e4.slice/crio-488adecb445c2bffe9da602932e78983aa1f5aab1bd56c8dee5635f5ac148552 WatchSource:0}: Error finding container 488adecb445c2bffe9da602932e78983aa1f5aab1bd56c8dee5635f5ac148552: Status 404 returned error can't find the container with id 488adecb445c2bffe9da602932e78983aa1f5aab1bd56c8dee5635f5ac148552 Dec 06 03:54:45 crc kubenswrapper[4980]: I1206 03:54:45.030943 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-sync-bdqlq" event={"ID":"d8db6f70-28bc-4b58-91e9-feb5bd06a3e4","Type":"ContainerStarted","Data":"488adecb445c2bffe9da602932e78983aa1f5aab1bd56c8dee5635f5ac148552"} Dec 06 03:55:04 crc kubenswrapper[4980]: E1206 03:55:04.017909 4980 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-glance-api:current-podified" Dec 06 03:55:04 crc kubenswrapper[4980]: E1206 03:55:04.018614 4980 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:glance-db-sync,Image:quay.io/podified-antelope-centos9/openstack-glance-api:current-podified,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:true,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/glance/glance.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/my.cnf,SubPath:my.cnf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:db-sync-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-5rh6z,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42415,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:*42415,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod glance-db-sync-bdqlq_glance-kuttl-tests(d8db6f70-28bc-4b58-91e9-feb5bd06a3e4): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 06 03:55:04 crc kubenswrapper[4980]: E1206 03:55:04.019791 4980 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"glance-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="glance-kuttl-tests/glance-db-sync-bdqlq" podUID="d8db6f70-28bc-4b58-91e9-feb5bd06a3e4" Dec 06 03:55:04 crc kubenswrapper[4980]: E1206 03:55:04.173411 4980 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"glance-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-glance-api:current-podified\\\"\"" pod="glance-kuttl-tests/glance-db-sync-bdqlq" podUID="d8db6f70-28bc-4b58-91e9-feb5bd06a3e4" Dec 06 03:55:16 crc kubenswrapper[4980]: I1206 03:55:16.327158 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-sync-bdqlq" event={"ID":"d8db6f70-28bc-4b58-91e9-feb5bd06a3e4","Type":"ContainerStarted","Data":"fdff1a97bc85df51ddd2b75327e3acbbe68356f8c40a00b7aaa611aac0ccbbc0"} Dec 06 03:55:16 crc kubenswrapper[4980]: I1206 03:55:16.349079 4980 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/glance-db-sync-bdqlq" podStartSLOduration=1.8452601180000001 podStartE2EDuration="33.349037474s" podCreationTimestamp="2025-12-06 03:54:43 +0000 UTC" firstStartedPulling="2025-12-06 03:54:44.202891625 +0000 UTC m=+1243.447642896" lastFinishedPulling="2025-12-06 03:55:15.706668981 +0000 UTC m=+1274.951420252" observedRunningTime="2025-12-06 03:55:16.346772949 +0000 UTC m=+1275.591524220" watchObservedRunningTime="2025-12-06 03:55:16.349037474 +0000 UTC m=+1275.593788745" Dec 06 03:55:23 crc kubenswrapper[4980]: I1206 03:55:23.395298 4980 generic.go:334] "Generic (PLEG): container finished" podID="d8db6f70-28bc-4b58-91e9-feb5bd06a3e4" containerID="fdff1a97bc85df51ddd2b75327e3acbbe68356f8c40a00b7aaa611aac0ccbbc0" exitCode=0 Dec 06 03:55:23 crc kubenswrapper[4980]: I1206 03:55:23.395404 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-sync-bdqlq" event={"ID":"d8db6f70-28bc-4b58-91e9-feb5bd06a3e4","Type":"ContainerDied","Data":"fdff1a97bc85df51ddd2b75327e3acbbe68356f8c40a00b7aaa611aac0ccbbc0"} Dec 06 03:55:24 crc kubenswrapper[4980]: I1206 03:55:24.700136 4980 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-sync-bdqlq" Dec 06 03:55:24 crc kubenswrapper[4980]: I1206 03:55:24.829865 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/d8db6f70-28bc-4b58-91e9-feb5bd06a3e4-db-sync-config-data\") pod \"d8db6f70-28bc-4b58-91e9-feb5bd06a3e4\" (UID: \"d8db6f70-28bc-4b58-91e9-feb5bd06a3e4\") " Dec 06 03:55:24 crc kubenswrapper[4980]: I1206 03:55:24.830014 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d8db6f70-28bc-4b58-91e9-feb5bd06a3e4-config-data\") pod \"d8db6f70-28bc-4b58-91e9-feb5bd06a3e4\" (UID: \"d8db6f70-28bc-4b58-91e9-feb5bd06a3e4\") " Dec 06 03:55:24 crc kubenswrapper[4980]: I1206 03:55:24.830111 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5rh6z\" (UniqueName: \"kubernetes.io/projected/d8db6f70-28bc-4b58-91e9-feb5bd06a3e4-kube-api-access-5rh6z\") pod \"d8db6f70-28bc-4b58-91e9-feb5bd06a3e4\" (UID: \"d8db6f70-28bc-4b58-91e9-feb5bd06a3e4\") " Dec 06 03:55:24 crc kubenswrapper[4980]: I1206 03:55:24.835787 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d8db6f70-28bc-4b58-91e9-feb5bd06a3e4-kube-api-access-5rh6z" (OuterVolumeSpecName: "kube-api-access-5rh6z") pod "d8db6f70-28bc-4b58-91e9-feb5bd06a3e4" (UID: "d8db6f70-28bc-4b58-91e9-feb5bd06a3e4"). InnerVolumeSpecName "kube-api-access-5rh6z". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 03:55:24 crc kubenswrapper[4980]: I1206 03:55:24.835869 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d8db6f70-28bc-4b58-91e9-feb5bd06a3e4-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "d8db6f70-28bc-4b58-91e9-feb5bd06a3e4" (UID: "d8db6f70-28bc-4b58-91e9-feb5bd06a3e4"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 03:55:24 crc kubenswrapper[4980]: I1206 03:55:24.877956 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d8db6f70-28bc-4b58-91e9-feb5bd06a3e4-config-data" (OuterVolumeSpecName: "config-data") pod "d8db6f70-28bc-4b58-91e9-feb5bd06a3e4" (UID: "d8db6f70-28bc-4b58-91e9-feb5bd06a3e4"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 03:55:24 crc kubenswrapper[4980]: I1206 03:55:24.932998 4980 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/d8db6f70-28bc-4b58-91e9-feb5bd06a3e4-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 03:55:24 crc kubenswrapper[4980]: I1206 03:55:24.933055 4980 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d8db6f70-28bc-4b58-91e9-feb5bd06a3e4-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 03:55:24 crc kubenswrapper[4980]: I1206 03:55:24.933077 4980 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5rh6z\" (UniqueName: \"kubernetes.io/projected/d8db6f70-28bc-4b58-91e9-feb5bd06a3e4-kube-api-access-5rh6z\") on node \"crc\" DevicePath \"\"" Dec 06 03:55:25 crc kubenswrapper[4980]: I1206 03:55:25.137532 4980 patch_prober.go:28] interesting pod/machine-config-daemon-r5zfb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 03:55:25 crc kubenswrapper[4980]: I1206 03:55:25.137652 4980 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-r5zfb" podUID="320f44d1-a671-4a91-b328-a8b0fdd8f23a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 03:55:25 crc kubenswrapper[4980]: I1206 03:55:25.426293 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-sync-bdqlq" event={"ID":"d8db6f70-28bc-4b58-91e9-feb5bd06a3e4","Type":"ContainerDied","Data":"488adecb445c2bffe9da602932e78983aa1f5aab1bd56c8dee5635f5ac148552"} Dec 06 03:55:25 crc kubenswrapper[4980]: I1206 03:55:25.426343 4980 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="488adecb445c2bffe9da602932e78983aa1f5aab1bd56c8dee5635f5ac148552" Dec 06 03:55:25 crc kubenswrapper[4980]: I1206 03:55:25.426376 4980 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-sync-bdqlq" Dec 06 03:55:26 crc kubenswrapper[4980]: I1206 03:55:26.803582 4980 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Dec 06 03:55:26 crc kubenswrapper[4980]: E1206 03:55:26.803896 4980 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d8db6f70-28bc-4b58-91e9-feb5bd06a3e4" containerName="glance-db-sync" Dec 06 03:55:26 crc kubenswrapper[4980]: I1206 03:55:26.803910 4980 state_mem.go:107] "Deleted CPUSet assignment" podUID="d8db6f70-28bc-4b58-91e9-feb5bd06a3e4" containerName="glance-db-sync" Dec 06 03:55:26 crc kubenswrapper[4980]: I1206 03:55:26.804041 4980 memory_manager.go:354] "RemoveStaleState removing state" podUID="d8db6f70-28bc-4b58-91e9-feb5bd06a3e4" containerName="glance-db-sync" Dec 06 03:55:26 crc kubenswrapper[4980]: I1206 03:55:26.804795 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-0" Dec 06 03:55:26 crc kubenswrapper[4980]: I1206 03:55:26.810862 4980 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-glance-dockercfg-kpxrp" Dec 06 03:55:26 crc kubenswrapper[4980]: I1206 03:55:26.810987 4980 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-default-single-config-data" Dec 06 03:55:26 crc kubenswrapper[4980]: I1206 03:55:26.811492 4980 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-scripts" Dec 06 03:55:26 crc kubenswrapper[4980]: I1206 03:55:26.828270 4980 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Dec 06 03:55:26 crc kubenswrapper[4980]: I1206 03:55:26.836281 4980 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-default-single-1"] Dec 06 03:55:26 crc kubenswrapper[4980]: I1206 03:55:26.837690 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-1" Dec 06 03:55:26 crc kubenswrapper[4980]: I1206 03:55:26.866979 4980 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-single-1"] Dec 06 03:55:26 crc kubenswrapper[4980]: I1206 03:55:26.963619 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/006f4ba2-a479-400f-99c6-468450cbe52f-etc-iscsi\") pod \"glance-default-single-1\" (UID: \"006f4ba2-a479-400f-99c6-468450cbe52f\") " pod="glance-kuttl-tests/glance-default-single-1" Dec 06 03:55:26 crc kubenswrapper[4980]: I1206 03:55:26.963679 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/006f4ba2-a479-400f-99c6-468450cbe52f-etc-nvme\") pod \"glance-default-single-1\" (UID: \"006f4ba2-a479-400f-99c6-468450cbe52f\") " pod="glance-kuttl-tests/glance-default-single-1" Dec 06 03:55:26 crc kubenswrapper[4980]: I1206 03:55:26.963732 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/006f4ba2-a479-400f-99c6-468450cbe52f-lib-modules\") pod \"glance-default-single-1\" (UID: \"006f4ba2-a479-400f-99c6-468450cbe52f\") " pod="glance-kuttl-tests/glance-default-single-1" Dec 06 03:55:26 crc kubenswrapper[4980]: I1206 03:55:26.963756 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/006f4ba2-a479-400f-99c6-468450cbe52f-logs\") pod \"glance-default-single-1\" (UID: \"006f4ba2-a479-400f-99c6-468450cbe52f\") " pod="glance-kuttl-tests/glance-default-single-1" Dec 06 03:55:26 crc kubenswrapper[4980]: I1206 03:55:26.963819 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/006f4ba2-a479-400f-99c6-468450cbe52f-run\") pod \"glance-default-single-1\" (UID: \"006f4ba2-a479-400f-99c6-468450cbe52f\") " pod="glance-kuttl-tests/glance-default-single-1" Dec 06 03:55:26 crc kubenswrapper[4980]: I1206 03:55:26.963839 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/de2e2787-7471-41a3-9bd7-e591410f3f42-sys\") pod \"glance-default-single-0\" (UID: \"de2e2787-7471-41a3-9bd7-e591410f3f42\") " pod="glance-kuttl-tests/glance-default-single-0" Dec 06 03:55:26 crc kubenswrapper[4980]: I1206 03:55:26.963865 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-single-0\" (UID: \"de2e2787-7471-41a3-9bd7-e591410f3f42\") " pod="glance-kuttl-tests/glance-default-single-0" Dec 06 03:55:26 crc kubenswrapper[4980]: I1206 03:55:26.963896 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-99bgs\" (UniqueName: \"kubernetes.io/projected/de2e2787-7471-41a3-9bd7-e591410f3f42-kube-api-access-99bgs\") pod \"glance-default-single-0\" (UID: \"de2e2787-7471-41a3-9bd7-e591410f3f42\") " pod="glance-kuttl-tests/glance-default-single-0" Dec 06 03:55:26 crc kubenswrapper[4980]: I1206 03:55:26.963915 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/006f4ba2-a479-400f-99c6-468450cbe52f-sys\") pod \"glance-default-single-1\" (UID: \"006f4ba2-a479-400f-99c6-468450cbe52f\") " pod="glance-kuttl-tests/glance-default-single-1" Dec 06 03:55:26 crc kubenswrapper[4980]: I1206 03:55:26.963934 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/de2e2787-7471-41a3-9bd7-e591410f3f42-httpd-run\") pod \"glance-default-single-0\" (UID: \"de2e2787-7471-41a3-9bd7-e591410f3f42\") " pod="glance-kuttl-tests/glance-default-single-0" Dec 06 03:55:26 crc kubenswrapper[4980]: I1206 03:55:26.963966 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/006f4ba2-a479-400f-99c6-468450cbe52f-scripts\") pod \"glance-default-single-1\" (UID: \"006f4ba2-a479-400f-99c6-468450cbe52f\") " pod="glance-kuttl-tests/glance-default-single-1" Dec 06 03:55:26 crc kubenswrapper[4980]: I1206 03:55:26.964004 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/006f4ba2-a479-400f-99c6-468450cbe52f-httpd-run\") pod \"glance-default-single-1\" (UID: \"006f4ba2-a479-400f-99c6-468450cbe52f\") " pod="glance-kuttl-tests/glance-default-single-1" Dec 06 03:55:26 crc kubenswrapper[4980]: I1206 03:55:26.964030 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/de2e2787-7471-41a3-9bd7-e591410f3f42-dev\") pod \"glance-default-single-0\" (UID: \"de2e2787-7471-41a3-9bd7-e591410f3f42\") " pod="glance-kuttl-tests/glance-default-single-0" Dec 06 03:55:26 crc kubenswrapper[4980]: I1206 03:55:26.964072 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/de2e2787-7471-41a3-9bd7-e591410f3f42-var-locks-brick\") pod \"glance-default-single-0\" (UID: \"de2e2787-7471-41a3-9bd7-e591410f3f42\") " pod="glance-kuttl-tests/glance-default-single-0" Dec 06 03:55:26 crc kubenswrapper[4980]: I1206 03:55:26.964101 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/de2e2787-7471-41a3-9bd7-e591410f3f42-lib-modules\") pod \"glance-default-single-0\" (UID: \"de2e2787-7471-41a3-9bd7-e591410f3f42\") " pod="glance-kuttl-tests/glance-default-single-0" Dec 06 03:55:26 crc kubenswrapper[4980]: I1206 03:55:26.964128 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/de2e2787-7471-41a3-9bd7-e591410f3f42-scripts\") pod \"glance-default-single-0\" (UID: \"de2e2787-7471-41a3-9bd7-e591410f3f42\") " pod="glance-kuttl-tests/glance-default-single-0" Dec 06 03:55:26 crc kubenswrapper[4980]: I1206 03:55:26.964203 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/de2e2787-7471-41a3-9bd7-e591410f3f42-logs\") pod \"glance-default-single-0\" (UID: \"de2e2787-7471-41a3-9bd7-e591410f3f42\") " pod="glance-kuttl-tests/glance-default-single-0" Dec 06 03:55:26 crc kubenswrapper[4980]: I1206 03:55:26.964269 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/006f4ba2-a479-400f-99c6-468450cbe52f-var-locks-brick\") pod \"glance-default-single-1\" (UID: \"006f4ba2-a479-400f-99c6-468450cbe52f\") " pod="glance-kuttl-tests/glance-default-single-1" Dec 06 03:55:26 crc kubenswrapper[4980]: I1206 03:55:26.964347 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-single-1\" (UID: \"006f4ba2-a479-400f-99c6-468450cbe52f\") " pod="glance-kuttl-tests/glance-default-single-1" Dec 06 03:55:26 crc kubenswrapper[4980]: I1206 03:55:26.964436 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/de2e2787-7471-41a3-9bd7-e591410f3f42-run\") pod \"glance-default-single-0\" (UID: \"de2e2787-7471-41a3-9bd7-e591410f3f42\") " pod="glance-kuttl-tests/glance-default-single-0" Dec 06 03:55:26 crc kubenswrapper[4980]: I1206 03:55:26.964497 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/de2e2787-7471-41a3-9bd7-e591410f3f42-etc-nvme\") pod \"glance-default-single-0\" (UID: \"de2e2787-7471-41a3-9bd7-e591410f3f42\") " pod="glance-kuttl-tests/glance-default-single-0" Dec 06 03:55:26 crc kubenswrapper[4980]: I1206 03:55:26.964582 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7bwrm\" (UniqueName: \"kubernetes.io/projected/006f4ba2-a479-400f-99c6-468450cbe52f-kube-api-access-7bwrm\") pod \"glance-default-single-1\" (UID: \"006f4ba2-a479-400f-99c6-468450cbe52f\") " pod="glance-kuttl-tests/glance-default-single-1" Dec 06 03:55:26 crc kubenswrapper[4980]: I1206 03:55:26.964628 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/de2e2787-7471-41a3-9bd7-e591410f3f42-config-data\") pod \"glance-default-single-0\" (UID: \"de2e2787-7471-41a3-9bd7-e591410f3f42\") " pod="glance-kuttl-tests/glance-default-single-0" Dec 06 03:55:26 crc kubenswrapper[4980]: I1206 03:55:26.964657 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-single-0\" (UID: \"de2e2787-7471-41a3-9bd7-e591410f3f42\") " pod="glance-kuttl-tests/glance-default-single-0" Dec 06 03:55:26 crc kubenswrapper[4980]: I1206 03:55:26.964677 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-single-1\" (UID: \"006f4ba2-a479-400f-99c6-468450cbe52f\") " pod="glance-kuttl-tests/glance-default-single-1" Dec 06 03:55:26 crc kubenswrapper[4980]: I1206 03:55:26.964835 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/de2e2787-7471-41a3-9bd7-e591410f3f42-etc-iscsi\") pod \"glance-default-single-0\" (UID: \"de2e2787-7471-41a3-9bd7-e591410f3f42\") " pod="glance-kuttl-tests/glance-default-single-0" Dec 06 03:55:26 crc kubenswrapper[4980]: I1206 03:55:26.964891 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/006f4ba2-a479-400f-99c6-468450cbe52f-config-data\") pod \"glance-default-single-1\" (UID: \"006f4ba2-a479-400f-99c6-468450cbe52f\") " pod="glance-kuttl-tests/glance-default-single-1" Dec 06 03:55:26 crc kubenswrapper[4980]: I1206 03:55:26.964908 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/006f4ba2-a479-400f-99c6-468450cbe52f-dev\") pod \"glance-default-single-1\" (UID: \"006f4ba2-a479-400f-99c6-468450cbe52f\") " pod="glance-kuttl-tests/glance-default-single-1" Dec 06 03:55:27 crc kubenswrapper[4980]: I1206 03:55:27.066270 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/006f4ba2-a479-400f-99c6-468450cbe52f-var-locks-brick\") pod \"glance-default-single-1\" (UID: \"006f4ba2-a479-400f-99c6-468450cbe52f\") " pod="glance-kuttl-tests/glance-default-single-1" Dec 06 03:55:27 crc kubenswrapper[4980]: I1206 03:55:27.066337 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-single-1\" (UID: \"006f4ba2-a479-400f-99c6-468450cbe52f\") " pod="glance-kuttl-tests/glance-default-single-1" Dec 06 03:55:27 crc kubenswrapper[4980]: I1206 03:55:27.066367 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/de2e2787-7471-41a3-9bd7-e591410f3f42-run\") pod \"glance-default-single-0\" (UID: \"de2e2787-7471-41a3-9bd7-e591410f3f42\") " pod="glance-kuttl-tests/glance-default-single-0" Dec 06 03:55:27 crc kubenswrapper[4980]: I1206 03:55:27.066383 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/de2e2787-7471-41a3-9bd7-e591410f3f42-etc-nvme\") pod \"glance-default-single-0\" (UID: \"de2e2787-7471-41a3-9bd7-e591410f3f42\") " pod="glance-kuttl-tests/glance-default-single-0" Dec 06 03:55:27 crc kubenswrapper[4980]: I1206 03:55:27.066413 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7bwrm\" (UniqueName: \"kubernetes.io/projected/006f4ba2-a479-400f-99c6-468450cbe52f-kube-api-access-7bwrm\") pod \"glance-default-single-1\" (UID: \"006f4ba2-a479-400f-99c6-468450cbe52f\") " pod="glance-kuttl-tests/glance-default-single-1" Dec 06 03:55:27 crc kubenswrapper[4980]: I1206 03:55:27.066432 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/de2e2787-7471-41a3-9bd7-e591410f3f42-config-data\") pod \"glance-default-single-0\" (UID: \"de2e2787-7471-41a3-9bd7-e591410f3f42\") " pod="glance-kuttl-tests/glance-default-single-0" Dec 06 03:55:27 crc kubenswrapper[4980]: I1206 03:55:27.066453 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-single-0\" (UID: \"de2e2787-7471-41a3-9bd7-e591410f3f42\") " pod="glance-kuttl-tests/glance-default-single-0" Dec 06 03:55:27 crc kubenswrapper[4980]: I1206 03:55:27.066467 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-single-1\" (UID: \"006f4ba2-a479-400f-99c6-468450cbe52f\") " pod="glance-kuttl-tests/glance-default-single-1" Dec 06 03:55:27 crc kubenswrapper[4980]: I1206 03:55:27.066493 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/de2e2787-7471-41a3-9bd7-e591410f3f42-etc-iscsi\") pod \"glance-default-single-0\" (UID: \"de2e2787-7471-41a3-9bd7-e591410f3f42\") " pod="glance-kuttl-tests/glance-default-single-0" Dec 06 03:55:27 crc kubenswrapper[4980]: I1206 03:55:27.066531 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/006f4ba2-a479-400f-99c6-468450cbe52f-config-data\") pod \"glance-default-single-1\" (UID: \"006f4ba2-a479-400f-99c6-468450cbe52f\") " pod="glance-kuttl-tests/glance-default-single-1" Dec 06 03:55:27 crc kubenswrapper[4980]: I1206 03:55:27.066548 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/006f4ba2-a479-400f-99c6-468450cbe52f-dev\") pod \"glance-default-single-1\" (UID: \"006f4ba2-a479-400f-99c6-468450cbe52f\") " pod="glance-kuttl-tests/glance-default-single-1" Dec 06 03:55:27 crc kubenswrapper[4980]: I1206 03:55:27.066563 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/006f4ba2-a479-400f-99c6-468450cbe52f-etc-nvme\") pod \"glance-default-single-1\" (UID: \"006f4ba2-a479-400f-99c6-468450cbe52f\") " pod="glance-kuttl-tests/glance-default-single-1" Dec 06 03:55:27 crc kubenswrapper[4980]: I1206 03:55:27.066578 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/006f4ba2-a479-400f-99c6-468450cbe52f-etc-iscsi\") pod \"glance-default-single-1\" (UID: \"006f4ba2-a479-400f-99c6-468450cbe52f\") " pod="glance-kuttl-tests/glance-default-single-1" Dec 06 03:55:27 crc kubenswrapper[4980]: I1206 03:55:27.066600 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/006f4ba2-a479-400f-99c6-468450cbe52f-logs\") pod \"glance-default-single-1\" (UID: \"006f4ba2-a479-400f-99c6-468450cbe52f\") " pod="glance-kuttl-tests/glance-default-single-1" Dec 06 03:55:27 crc kubenswrapper[4980]: I1206 03:55:27.066615 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/006f4ba2-a479-400f-99c6-468450cbe52f-lib-modules\") pod \"glance-default-single-1\" (UID: \"006f4ba2-a479-400f-99c6-468450cbe52f\") " pod="glance-kuttl-tests/glance-default-single-1" Dec 06 03:55:27 crc kubenswrapper[4980]: I1206 03:55:27.066632 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/006f4ba2-a479-400f-99c6-468450cbe52f-run\") pod \"glance-default-single-1\" (UID: \"006f4ba2-a479-400f-99c6-468450cbe52f\") " pod="glance-kuttl-tests/glance-default-single-1" Dec 06 03:55:27 crc kubenswrapper[4980]: I1206 03:55:27.066646 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/de2e2787-7471-41a3-9bd7-e591410f3f42-sys\") pod \"glance-default-single-0\" (UID: \"de2e2787-7471-41a3-9bd7-e591410f3f42\") " pod="glance-kuttl-tests/glance-default-single-0" Dec 06 03:55:27 crc kubenswrapper[4980]: I1206 03:55:27.066665 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-single-0\" (UID: \"de2e2787-7471-41a3-9bd7-e591410f3f42\") " pod="glance-kuttl-tests/glance-default-single-0" Dec 06 03:55:27 crc kubenswrapper[4980]: I1206 03:55:27.066685 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-99bgs\" (UniqueName: \"kubernetes.io/projected/de2e2787-7471-41a3-9bd7-e591410f3f42-kube-api-access-99bgs\") pod \"glance-default-single-0\" (UID: \"de2e2787-7471-41a3-9bd7-e591410f3f42\") " pod="glance-kuttl-tests/glance-default-single-0" Dec 06 03:55:27 crc kubenswrapper[4980]: I1206 03:55:27.066701 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/006f4ba2-a479-400f-99c6-468450cbe52f-sys\") pod \"glance-default-single-1\" (UID: \"006f4ba2-a479-400f-99c6-468450cbe52f\") " pod="glance-kuttl-tests/glance-default-single-1" Dec 06 03:55:27 crc kubenswrapper[4980]: I1206 03:55:27.066719 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/de2e2787-7471-41a3-9bd7-e591410f3f42-httpd-run\") pod \"glance-default-single-0\" (UID: \"de2e2787-7471-41a3-9bd7-e591410f3f42\") " pod="glance-kuttl-tests/glance-default-single-0" Dec 06 03:55:27 crc kubenswrapper[4980]: I1206 03:55:27.066734 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/006f4ba2-a479-400f-99c6-468450cbe52f-scripts\") pod \"glance-default-single-1\" (UID: \"006f4ba2-a479-400f-99c6-468450cbe52f\") " pod="glance-kuttl-tests/glance-default-single-1" Dec 06 03:55:27 crc kubenswrapper[4980]: I1206 03:55:27.066760 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/006f4ba2-a479-400f-99c6-468450cbe52f-httpd-run\") pod \"glance-default-single-1\" (UID: \"006f4ba2-a479-400f-99c6-468450cbe52f\") " pod="glance-kuttl-tests/glance-default-single-1" Dec 06 03:55:27 crc kubenswrapper[4980]: I1206 03:55:27.066762 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/de2e2787-7471-41a3-9bd7-e591410f3f42-etc-nvme\") pod \"glance-default-single-0\" (UID: \"de2e2787-7471-41a3-9bd7-e591410f3f42\") " pod="glance-kuttl-tests/glance-default-single-0" Dec 06 03:55:27 crc kubenswrapper[4980]: I1206 03:55:27.066782 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/de2e2787-7471-41a3-9bd7-e591410f3f42-dev\") pod \"glance-default-single-0\" (UID: \"de2e2787-7471-41a3-9bd7-e591410f3f42\") " pod="glance-kuttl-tests/glance-default-single-0" Dec 06 03:55:27 crc kubenswrapper[4980]: I1206 03:55:27.066797 4980 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-single-1\" (UID: \"006f4ba2-a479-400f-99c6-468450cbe52f\") device mount path \"/mnt/openstack/pv08\"" pod="glance-kuttl-tests/glance-default-single-1" Dec 06 03:55:27 crc kubenswrapper[4980]: I1206 03:55:27.066463 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/de2e2787-7471-41a3-9bd7-e591410f3f42-run\") pod \"glance-default-single-0\" (UID: \"de2e2787-7471-41a3-9bd7-e591410f3f42\") " pod="glance-kuttl-tests/glance-default-single-0" Dec 06 03:55:27 crc kubenswrapper[4980]: I1206 03:55:27.066825 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/de2e2787-7471-41a3-9bd7-e591410f3f42-dev\") pod \"glance-default-single-0\" (UID: \"de2e2787-7471-41a3-9bd7-e591410f3f42\") " pod="glance-kuttl-tests/glance-default-single-0" Dec 06 03:55:27 crc kubenswrapper[4980]: I1206 03:55:27.066826 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/006f4ba2-a479-400f-99c6-468450cbe52f-var-locks-brick\") pod \"glance-default-single-1\" (UID: \"006f4ba2-a479-400f-99c6-468450cbe52f\") " pod="glance-kuttl-tests/glance-default-single-1" Dec 06 03:55:27 crc kubenswrapper[4980]: I1206 03:55:27.066868 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/006f4ba2-a479-400f-99c6-468450cbe52f-run\") pod \"glance-default-single-1\" (UID: \"006f4ba2-a479-400f-99c6-468450cbe52f\") " pod="glance-kuttl-tests/glance-default-single-1" Dec 06 03:55:27 crc kubenswrapper[4980]: I1206 03:55:27.066896 4980 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-single-1\" (UID: \"006f4ba2-a479-400f-99c6-468450cbe52f\") device mount path \"/mnt/openstack/pv07\"" pod="glance-kuttl-tests/glance-default-single-1" Dec 06 03:55:27 crc kubenswrapper[4980]: I1206 03:55:27.066908 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/de2e2787-7471-41a3-9bd7-e591410f3f42-sys\") pod \"glance-default-single-0\" (UID: \"de2e2787-7471-41a3-9bd7-e591410f3f42\") " pod="glance-kuttl-tests/glance-default-single-0" Dec 06 03:55:27 crc kubenswrapper[4980]: I1206 03:55:27.066907 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/006f4ba2-a479-400f-99c6-468450cbe52f-etc-nvme\") pod \"glance-default-single-1\" (UID: \"006f4ba2-a479-400f-99c6-468450cbe52f\") " pod="glance-kuttl-tests/glance-default-single-1" Dec 06 03:55:27 crc kubenswrapper[4980]: I1206 03:55:27.066939 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/006f4ba2-a479-400f-99c6-468450cbe52f-dev\") pod \"glance-default-single-1\" (UID: \"006f4ba2-a479-400f-99c6-468450cbe52f\") " pod="glance-kuttl-tests/glance-default-single-1" Dec 06 03:55:27 crc kubenswrapper[4980]: I1206 03:55:27.067014 4980 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-single-0\" (UID: \"de2e2787-7471-41a3-9bd7-e591410f3f42\") device mount path \"/mnt/openstack/pv05\"" pod="glance-kuttl-tests/glance-default-single-0" Dec 06 03:55:27 crc kubenswrapper[4980]: I1206 03:55:27.067018 4980 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-single-0\" (UID: \"de2e2787-7471-41a3-9bd7-e591410f3f42\") device mount path \"/mnt/openstack/pv09\"" pod="glance-kuttl-tests/glance-default-single-0" Dec 06 03:55:27 crc kubenswrapper[4980]: I1206 03:55:27.067038 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/006f4ba2-a479-400f-99c6-468450cbe52f-etc-iscsi\") pod \"glance-default-single-1\" (UID: \"006f4ba2-a479-400f-99c6-468450cbe52f\") " pod="glance-kuttl-tests/glance-default-single-1" Dec 06 03:55:27 crc kubenswrapper[4980]: I1206 03:55:27.067091 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/006f4ba2-a479-400f-99c6-468450cbe52f-sys\") pod \"glance-default-single-1\" (UID: \"006f4ba2-a479-400f-99c6-468450cbe52f\") " pod="glance-kuttl-tests/glance-default-single-1" Dec 06 03:55:27 crc kubenswrapper[4980]: I1206 03:55:27.067175 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/de2e2787-7471-41a3-9bd7-e591410f3f42-var-locks-brick\") pod \"glance-default-single-0\" (UID: \"de2e2787-7471-41a3-9bd7-e591410f3f42\") " pod="glance-kuttl-tests/glance-default-single-0" Dec 06 03:55:27 crc kubenswrapper[4980]: I1206 03:55:27.067220 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/de2e2787-7471-41a3-9bd7-e591410f3f42-etc-iscsi\") pod \"glance-default-single-0\" (UID: \"de2e2787-7471-41a3-9bd7-e591410f3f42\") " pod="glance-kuttl-tests/glance-default-single-0" Dec 06 03:55:27 crc kubenswrapper[4980]: I1206 03:55:27.067261 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/de2e2787-7471-41a3-9bd7-e591410f3f42-var-locks-brick\") pod \"glance-default-single-0\" (UID: \"de2e2787-7471-41a3-9bd7-e591410f3f42\") " pod="glance-kuttl-tests/glance-default-single-0" Dec 06 03:55:27 crc kubenswrapper[4980]: I1206 03:55:27.067258 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/006f4ba2-a479-400f-99c6-468450cbe52f-lib-modules\") pod \"glance-default-single-1\" (UID: \"006f4ba2-a479-400f-99c6-468450cbe52f\") " pod="glance-kuttl-tests/glance-default-single-1" Dec 06 03:55:27 crc kubenswrapper[4980]: I1206 03:55:27.067301 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/de2e2787-7471-41a3-9bd7-e591410f3f42-lib-modules\") pod \"glance-default-single-0\" (UID: \"de2e2787-7471-41a3-9bd7-e591410f3f42\") " pod="glance-kuttl-tests/glance-default-single-0" Dec 06 03:55:27 crc kubenswrapper[4980]: I1206 03:55:27.067324 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/de2e2787-7471-41a3-9bd7-e591410f3f42-scripts\") pod \"glance-default-single-0\" (UID: \"de2e2787-7471-41a3-9bd7-e591410f3f42\") " pod="glance-kuttl-tests/glance-default-single-0" Dec 06 03:55:27 crc kubenswrapper[4980]: I1206 03:55:27.067367 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/de2e2787-7471-41a3-9bd7-e591410f3f42-lib-modules\") pod \"glance-default-single-0\" (UID: \"de2e2787-7471-41a3-9bd7-e591410f3f42\") " pod="glance-kuttl-tests/glance-default-single-0" Dec 06 03:55:27 crc kubenswrapper[4980]: I1206 03:55:27.067403 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/de2e2787-7471-41a3-9bd7-e591410f3f42-logs\") pod \"glance-default-single-0\" (UID: \"de2e2787-7471-41a3-9bd7-e591410f3f42\") " pod="glance-kuttl-tests/glance-default-single-0" Dec 06 03:55:27 crc kubenswrapper[4980]: I1206 03:55:27.067422 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/006f4ba2-a479-400f-99c6-468450cbe52f-httpd-run\") pod \"glance-default-single-1\" (UID: \"006f4ba2-a479-400f-99c6-468450cbe52f\") " pod="glance-kuttl-tests/glance-default-single-1" Dec 06 03:55:27 crc kubenswrapper[4980]: I1206 03:55:27.067462 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/de2e2787-7471-41a3-9bd7-e591410f3f42-httpd-run\") pod \"glance-default-single-0\" (UID: \"de2e2787-7471-41a3-9bd7-e591410f3f42\") " pod="glance-kuttl-tests/glance-default-single-0" Dec 06 03:55:27 crc kubenswrapper[4980]: I1206 03:55:27.068471 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/de2e2787-7471-41a3-9bd7-e591410f3f42-logs\") pod \"glance-default-single-0\" (UID: \"de2e2787-7471-41a3-9bd7-e591410f3f42\") " pod="glance-kuttl-tests/glance-default-single-0" Dec 06 03:55:27 crc kubenswrapper[4980]: I1206 03:55:27.068500 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/006f4ba2-a479-400f-99c6-468450cbe52f-logs\") pod \"glance-default-single-1\" (UID: \"006f4ba2-a479-400f-99c6-468450cbe52f\") " pod="glance-kuttl-tests/glance-default-single-1" Dec 06 03:55:27 crc kubenswrapper[4980]: I1206 03:55:27.080400 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/006f4ba2-a479-400f-99c6-468450cbe52f-config-data\") pod \"glance-default-single-1\" (UID: \"006f4ba2-a479-400f-99c6-468450cbe52f\") " pod="glance-kuttl-tests/glance-default-single-1" Dec 06 03:55:27 crc kubenswrapper[4980]: I1206 03:55:27.081655 4980 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-single-1"] Dec 06 03:55:27 crc kubenswrapper[4980]: E1206 03:55:27.082293 4980 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[glance glance-cache kube-api-access-7bwrm scripts], unattached volumes=[], failed to process volumes=[]: context canceled" pod="glance-kuttl-tests/glance-default-single-1" podUID="006f4ba2-a479-400f-99c6-468450cbe52f" Dec 06 03:55:27 crc kubenswrapper[4980]: I1206 03:55:27.086178 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/de2e2787-7471-41a3-9bd7-e591410f3f42-scripts\") pod \"glance-default-single-0\" (UID: \"de2e2787-7471-41a3-9bd7-e591410f3f42\") " pod="glance-kuttl-tests/glance-default-single-0" Dec 06 03:55:27 crc kubenswrapper[4980]: I1206 03:55:27.089376 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/006f4ba2-a479-400f-99c6-468450cbe52f-scripts\") pod \"glance-default-single-1\" (UID: \"006f4ba2-a479-400f-99c6-468450cbe52f\") " pod="glance-kuttl-tests/glance-default-single-1" Dec 06 03:55:27 crc kubenswrapper[4980]: I1206 03:55:27.089741 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/de2e2787-7471-41a3-9bd7-e591410f3f42-config-data\") pod \"glance-default-single-0\" (UID: \"de2e2787-7471-41a3-9bd7-e591410f3f42\") " pod="glance-kuttl-tests/glance-default-single-0" Dec 06 03:55:27 crc kubenswrapper[4980]: I1206 03:55:27.092700 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7bwrm\" (UniqueName: \"kubernetes.io/projected/006f4ba2-a479-400f-99c6-468450cbe52f-kube-api-access-7bwrm\") pod \"glance-default-single-1\" (UID: \"006f4ba2-a479-400f-99c6-468450cbe52f\") " pod="glance-kuttl-tests/glance-default-single-1" Dec 06 03:55:27 crc kubenswrapper[4980]: I1206 03:55:27.104269 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-99bgs\" (UniqueName: \"kubernetes.io/projected/de2e2787-7471-41a3-9bd7-e591410f3f42-kube-api-access-99bgs\") pod \"glance-default-single-0\" (UID: \"de2e2787-7471-41a3-9bd7-e591410f3f42\") " pod="glance-kuttl-tests/glance-default-single-0" Dec 06 03:55:27 crc kubenswrapper[4980]: I1206 03:55:27.119967 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-single-1\" (UID: \"006f4ba2-a479-400f-99c6-468450cbe52f\") " pod="glance-kuttl-tests/glance-default-single-1" Dec 06 03:55:27 crc kubenswrapper[4980]: I1206 03:55:27.121044 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-single-1\" (UID: \"006f4ba2-a479-400f-99c6-468450cbe52f\") " pod="glance-kuttl-tests/glance-default-single-1" Dec 06 03:55:27 crc kubenswrapper[4980]: I1206 03:55:27.130052 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-single-0\" (UID: \"de2e2787-7471-41a3-9bd7-e591410f3f42\") " pod="glance-kuttl-tests/glance-default-single-0" Dec 06 03:55:27 crc kubenswrapper[4980]: I1206 03:55:27.147921 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-single-0\" (UID: \"de2e2787-7471-41a3-9bd7-e591410f3f42\") " pod="glance-kuttl-tests/glance-default-single-0" Dec 06 03:55:27 crc kubenswrapper[4980]: I1206 03:55:27.424014 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-0" Dec 06 03:55:27 crc kubenswrapper[4980]: I1206 03:55:27.440066 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-1" Dec 06 03:55:27 crc kubenswrapper[4980]: I1206 03:55:27.504291 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-1" Dec 06 03:55:27 crc kubenswrapper[4980]: I1206 03:55:27.574501 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/006f4ba2-a479-400f-99c6-468450cbe52f-sys\") pod \"006f4ba2-a479-400f-99c6-468450cbe52f\" (UID: \"006f4ba2-a479-400f-99c6-468450cbe52f\") " Dec 06 03:55:27 crc kubenswrapper[4980]: I1206 03:55:27.574573 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/006f4ba2-a479-400f-99c6-468450cbe52f-httpd-run\") pod \"006f4ba2-a479-400f-99c6-468450cbe52f\" (UID: \"006f4ba2-a479-400f-99c6-468450cbe52f\") " Dec 06 03:55:27 crc kubenswrapper[4980]: I1206 03:55:27.574595 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/006f4ba2-a479-400f-99c6-468450cbe52f-var-locks-brick\") pod \"006f4ba2-a479-400f-99c6-468450cbe52f\" (UID: \"006f4ba2-a479-400f-99c6-468450cbe52f\") " Dec 06 03:55:27 crc kubenswrapper[4980]: I1206 03:55:27.574625 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"006f4ba2-a479-400f-99c6-468450cbe52f\" (UID: \"006f4ba2-a479-400f-99c6-468450cbe52f\") " Dec 06 03:55:27 crc kubenswrapper[4980]: I1206 03:55:27.574650 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance-cache\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"006f4ba2-a479-400f-99c6-468450cbe52f\" (UID: \"006f4ba2-a479-400f-99c6-468450cbe52f\") " Dec 06 03:55:27 crc kubenswrapper[4980]: I1206 03:55:27.574682 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/006f4ba2-a479-400f-99c6-468450cbe52f-config-data\") pod \"006f4ba2-a479-400f-99c6-468450cbe52f\" (UID: \"006f4ba2-a479-400f-99c6-468450cbe52f\") " Dec 06 03:55:27 crc kubenswrapper[4980]: I1206 03:55:27.574719 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/006f4ba2-a479-400f-99c6-468450cbe52f-logs\") pod \"006f4ba2-a479-400f-99c6-468450cbe52f\" (UID: \"006f4ba2-a479-400f-99c6-468450cbe52f\") " Dec 06 03:55:27 crc kubenswrapper[4980]: I1206 03:55:27.574763 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7bwrm\" (UniqueName: \"kubernetes.io/projected/006f4ba2-a479-400f-99c6-468450cbe52f-kube-api-access-7bwrm\") pod \"006f4ba2-a479-400f-99c6-468450cbe52f\" (UID: \"006f4ba2-a479-400f-99c6-468450cbe52f\") " Dec 06 03:55:27 crc kubenswrapper[4980]: I1206 03:55:27.574803 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/006f4ba2-a479-400f-99c6-468450cbe52f-etc-iscsi\") pod \"006f4ba2-a479-400f-99c6-468450cbe52f\" (UID: \"006f4ba2-a479-400f-99c6-468450cbe52f\") " Dec 06 03:55:27 crc kubenswrapper[4980]: I1206 03:55:27.574830 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/006f4ba2-a479-400f-99c6-468450cbe52f-lib-modules\") pod \"006f4ba2-a479-400f-99c6-468450cbe52f\" (UID: \"006f4ba2-a479-400f-99c6-468450cbe52f\") " Dec 06 03:55:27 crc kubenswrapper[4980]: I1206 03:55:27.574848 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/006f4ba2-a479-400f-99c6-468450cbe52f-scripts\") pod \"006f4ba2-a479-400f-99c6-468450cbe52f\" (UID: \"006f4ba2-a479-400f-99c6-468450cbe52f\") " Dec 06 03:55:27 crc kubenswrapper[4980]: I1206 03:55:27.574869 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/006f4ba2-a479-400f-99c6-468450cbe52f-etc-nvme\") pod \"006f4ba2-a479-400f-99c6-468450cbe52f\" (UID: \"006f4ba2-a479-400f-99c6-468450cbe52f\") " Dec 06 03:55:27 crc kubenswrapper[4980]: I1206 03:55:27.574919 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/006f4ba2-a479-400f-99c6-468450cbe52f-run\") pod \"006f4ba2-a479-400f-99c6-468450cbe52f\" (UID: \"006f4ba2-a479-400f-99c6-468450cbe52f\") " Dec 06 03:55:27 crc kubenswrapper[4980]: I1206 03:55:27.574977 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/006f4ba2-a479-400f-99c6-468450cbe52f-dev\") pod \"006f4ba2-a479-400f-99c6-468450cbe52f\" (UID: \"006f4ba2-a479-400f-99c6-468450cbe52f\") " Dec 06 03:55:27 crc kubenswrapper[4980]: I1206 03:55:27.575319 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/006f4ba2-a479-400f-99c6-468450cbe52f-dev" (OuterVolumeSpecName: "dev") pod "006f4ba2-a479-400f-99c6-468450cbe52f" (UID: "006f4ba2-a479-400f-99c6-468450cbe52f"). InnerVolumeSpecName "dev". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 03:55:27 crc kubenswrapper[4980]: I1206 03:55:27.575350 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/006f4ba2-a479-400f-99c6-468450cbe52f-sys" (OuterVolumeSpecName: "sys") pod "006f4ba2-a479-400f-99c6-468450cbe52f" (UID: "006f4ba2-a479-400f-99c6-468450cbe52f"). InnerVolumeSpecName "sys". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 03:55:27 crc kubenswrapper[4980]: I1206 03:55:27.575674 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/006f4ba2-a479-400f-99c6-468450cbe52f-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "006f4ba2-a479-400f-99c6-468450cbe52f" (UID: "006f4ba2-a479-400f-99c6-468450cbe52f"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 03:55:27 crc kubenswrapper[4980]: I1206 03:55:27.575748 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/006f4ba2-a479-400f-99c6-468450cbe52f-var-locks-brick" (OuterVolumeSpecName: "var-locks-brick") pod "006f4ba2-a479-400f-99c6-468450cbe52f" (UID: "006f4ba2-a479-400f-99c6-468450cbe52f"). InnerVolumeSpecName "var-locks-brick". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 03:55:27 crc kubenswrapper[4980]: I1206 03:55:27.588637 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/006f4ba2-a479-400f-99c6-468450cbe52f-etc-iscsi" (OuterVolumeSpecName: "etc-iscsi") pod "006f4ba2-a479-400f-99c6-468450cbe52f" (UID: "006f4ba2-a479-400f-99c6-468450cbe52f"). InnerVolumeSpecName "etc-iscsi". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 03:55:27 crc kubenswrapper[4980]: I1206 03:55:27.590215 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/006f4ba2-a479-400f-99c6-468450cbe52f-lib-modules" (OuterVolumeSpecName: "lib-modules") pod "006f4ba2-a479-400f-99c6-468450cbe52f" (UID: "006f4ba2-a479-400f-99c6-468450cbe52f"). InnerVolumeSpecName "lib-modules". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 03:55:27 crc kubenswrapper[4980]: I1206 03:55:27.590590 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/006f4ba2-a479-400f-99c6-468450cbe52f-etc-nvme" (OuterVolumeSpecName: "etc-nvme") pod "006f4ba2-a479-400f-99c6-468450cbe52f" (UID: "006f4ba2-a479-400f-99c6-468450cbe52f"). InnerVolumeSpecName "etc-nvme". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 03:55:27 crc kubenswrapper[4980]: I1206 03:55:27.590636 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/006f4ba2-a479-400f-99c6-468450cbe52f-run" (OuterVolumeSpecName: "run") pod "006f4ba2-a479-400f-99c6-468450cbe52f" (UID: "006f4ba2-a479-400f-99c6-468450cbe52f"). InnerVolumeSpecName "run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 03:55:27 crc kubenswrapper[4980]: I1206 03:55:27.590696 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage08-crc" (OuterVolumeSpecName: "glance") pod "006f4ba2-a479-400f-99c6-468450cbe52f" (UID: "006f4ba2-a479-400f-99c6-468450cbe52f"). InnerVolumeSpecName "local-storage08-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 06 03:55:27 crc kubenswrapper[4980]: I1206 03:55:27.590805 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/006f4ba2-a479-400f-99c6-468450cbe52f-logs" (OuterVolumeSpecName: "logs") pod "006f4ba2-a479-400f-99c6-468450cbe52f" (UID: "006f4ba2-a479-400f-99c6-468450cbe52f"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 03:55:27 crc kubenswrapper[4980]: I1206 03:55:27.592842 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage07-crc" (OuterVolumeSpecName: "glance-cache") pod "006f4ba2-a479-400f-99c6-468450cbe52f" (UID: "006f4ba2-a479-400f-99c6-468450cbe52f"). InnerVolumeSpecName "local-storage07-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 06 03:55:27 crc kubenswrapper[4980]: I1206 03:55:27.593842 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/006f4ba2-a479-400f-99c6-468450cbe52f-scripts" (OuterVolumeSpecName: "scripts") pod "006f4ba2-a479-400f-99c6-468450cbe52f" (UID: "006f4ba2-a479-400f-99c6-468450cbe52f"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 03:55:27 crc kubenswrapper[4980]: I1206 03:55:27.594033 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/006f4ba2-a479-400f-99c6-468450cbe52f-config-data" (OuterVolumeSpecName: "config-data") pod "006f4ba2-a479-400f-99c6-468450cbe52f" (UID: "006f4ba2-a479-400f-99c6-468450cbe52f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 03:55:27 crc kubenswrapper[4980]: I1206 03:55:27.594273 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/006f4ba2-a479-400f-99c6-468450cbe52f-kube-api-access-7bwrm" (OuterVolumeSpecName: "kube-api-access-7bwrm") pod "006f4ba2-a479-400f-99c6-468450cbe52f" (UID: "006f4ba2-a479-400f-99c6-468450cbe52f"). InnerVolumeSpecName "kube-api-access-7bwrm". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 03:55:27 crc kubenswrapper[4980]: I1206 03:55:27.677144 4980 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/006f4ba2-a479-400f-99c6-468450cbe52f-httpd-run\") on node \"crc\" DevicePath \"\"" Dec 06 03:55:27 crc kubenswrapper[4980]: I1206 03:55:27.677176 4980 reconciler_common.go:293] "Volume detached for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/006f4ba2-a479-400f-99c6-468450cbe52f-var-locks-brick\") on node \"crc\" DevicePath \"\"" Dec 06 03:55:27 crc kubenswrapper[4980]: I1206 03:55:27.677219 4980 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") on node \"crc\" " Dec 06 03:55:27 crc kubenswrapper[4980]: I1206 03:55:27.677235 4980 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") on node \"crc\" " Dec 06 03:55:27 crc kubenswrapper[4980]: I1206 03:55:27.677248 4980 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/006f4ba2-a479-400f-99c6-468450cbe52f-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 03:55:27 crc kubenswrapper[4980]: I1206 03:55:27.677262 4980 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/006f4ba2-a479-400f-99c6-468450cbe52f-logs\") on node \"crc\" DevicePath \"\"" Dec 06 03:55:27 crc kubenswrapper[4980]: I1206 03:55:27.677274 4980 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7bwrm\" (UniqueName: \"kubernetes.io/projected/006f4ba2-a479-400f-99c6-468450cbe52f-kube-api-access-7bwrm\") on node \"crc\" DevicePath \"\"" Dec 06 03:55:27 crc kubenswrapper[4980]: I1206 03:55:27.677286 4980 reconciler_common.go:293] "Volume detached for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/006f4ba2-a479-400f-99c6-468450cbe52f-etc-iscsi\") on node \"crc\" DevicePath \"\"" Dec 06 03:55:27 crc kubenswrapper[4980]: I1206 03:55:27.677297 4980 reconciler_common.go:293] "Volume detached for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/006f4ba2-a479-400f-99c6-468450cbe52f-lib-modules\") on node \"crc\" DevicePath \"\"" Dec 06 03:55:27 crc kubenswrapper[4980]: I1206 03:55:27.677308 4980 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/006f4ba2-a479-400f-99c6-468450cbe52f-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 03:55:27 crc kubenswrapper[4980]: I1206 03:55:27.677319 4980 reconciler_common.go:293] "Volume detached for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/006f4ba2-a479-400f-99c6-468450cbe52f-etc-nvme\") on node \"crc\" DevicePath \"\"" Dec 06 03:55:27 crc kubenswrapper[4980]: I1206 03:55:27.677330 4980 reconciler_common.go:293] "Volume detached for volume \"run\" (UniqueName: \"kubernetes.io/host-path/006f4ba2-a479-400f-99c6-468450cbe52f-run\") on node \"crc\" DevicePath \"\"" Dec 06 03:55:27 crc kubenswrapper[4980]: I1206 03:55:27.677342 4980 reconciler_common.go:293] "Volume detached for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/006f4ba2-a479-400f-99c6-468450cbe52f-dev\") on node \"crc\" DevicePath \"\"" Dec 06 03:55:27 crc kubenswrapper[4980]: I1206 03:55:27.677353 4980 reconciler_common.go:293] "Volume detached for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/006f4ba2-a479-400f-99c6-468450cbe52f-sys\") on node \"crc\" DevicePath \"\"" Dec 06 03:55:27 crc kubenswrapper[4980]: I1206 03:55:27.685784 4980 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Dec 06 03:55:27 crc kubenswrapper[4980]: I1206 03:55:27.691695 4980 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage08-crc" (UniqueName: "kubernetes.io/local-volume/local-storage08-crc") on node "crc" Dec 06 03:55:27 crc kubenswrapper[4980]: I1206 03:55:27.695544 4980 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage07-crc" (UniqueName: "kubernetes.io/local-volume/local-storage07-crc") on node "crc" Dec 06 03:55:27 crc kubenswrapper[4980]: I1206 03:55:27.779230 4980 reconciler_common.go:293] "Volume detached for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") on node \"crc\" DevicePath \"\"" Dec 06 03:55:27 crc kubenswrapper[4980]: I1206 03:55:27.779266 4980 reconciler_common.go:293] "Volume detached for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") on node \"crc\" DevicePath \"\"" Dec 06 03:55:28 crc kubenswrapper[4980]: I1206 03:55:28.451371 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"de2e2787-7471-41a3-9bd7-e591410f3f42","Type":"ContainerStarted","Data":"1432679ad8686abee0ae98476311d7a88fc012bdb4aeb04483111362fc779aa6"} Dec 06 03:55:28 crc kubenswrapper[4980]: I1206 03:55:28.451870 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"de2e2787-7471-41a3-9bd7-e591410f3f42","Type":"ContainerStarted","Data":"a98388d0fc55b4344bba2f679d8609c5a5c9f6019d03a5ae0d2772d41450746f"} Dec 06 03:55:28 crc kubenswrapper[4980]: I1206 03:55:28.451884 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"de2e2787-7471-41a3-9bd7-e591410f3f42","Type":"ContainerStarted","Data":"74f970886f6dbee3a5ce01e19cc473fac456170f227e1a7c881088af281cabd7"} Dec 06 03:55:28 crc kubenswrapper[4980]: I1206 03:55:28.451399 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-1" Dec 06 03:55:28 crc kubenswrapper[4980]: I1206 03:55:28.485057 4980 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/glance-default-single-0" podStartSLOduration=3.485036586 podStartE2EDuration="3.485036586s" podCreationTimestamp="2025-12-06 03:55:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 03:55:28.482657258 +0000 UTC m=+1287.727408529" watchObservedRunningTime="2025-12-06 03:55:28.485036586 +0000 UTC m=+1287.729787847" Dec 06 03:55:28 crc kubenswrapper[4980]: I1206 03:55:28.532212 4980 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-single-1"] Dec 06 03:55:28 crc kubenswrapper[4980]: I1206 03:55:28.532286 4980 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-default-single-1"] Dec 06 03:55:28 crc kubenswrapper[4980]: I1206 03:55:28.603990 4980 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-default-single-1"] Dec 06 03:55:28 crc kubenswrapper[4980]: I1206 03:55:28.605439 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-1" Dec 06 03:55:28 crc kubenswrapper[4980]: I1206 03:55:28.620497 4980 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-single-1"] Dec 06 03:55:28 crc kubenswrapper[4980]: I1206 03:55:28.792971 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/5bfed592-a73b-4218-bd4a-eab04765a748-lib-modules\") pod \"glance-default-single-1\" (UID: \"5bfed592-a73b-4218-bd4a-eab04765a748\") " pod="glance-kuttl-tests/glance-default-single-1" Dec 06 03:55:28 crc kubenswrapper[4980]: I1206 03:55:28.793016 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tvc5b\" (UniqueName: \"kubernetes.io/projected/5bfed592-a73b-4218-bd4a-eab04765a748-kube-api-access-tvc5b\") pod \"glance-default-single-1\" (UID: \"5bfed592-a73b-4218-bd4a-eab04765a748\") " pod="glance-kuttl-tests/glance-default-single-1" Dec 06 03:55:28 crc kubenswrapper[4980]: I1206 03:55:28.793052 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/5bfed592-a73b-4218-bd4a-eab04765a748-run\") pod \"glance-default-single-1\" (UID: \"5bfed592-a73b-4218-bd4a-eab04765a748\") " pod="glance-kuttl-tests/glance-default-single-1" Dec 06 03:55:28 crc kubenswrapper[4980]: I1206 03:55:28.793069 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/5bfed592-a73b-4218-bd4a-eab04765a748-etc-nvme\") pod \"glance-default-single-1\" (UID: \"5bfed592-a73b-4218-bd4a-eab04765a748\") " pod="glance-kuttl-tests/glance-default-single-1" Dec 06 03:55:28 crc kubenswrapper[4980]: I1206 03:55:28.793103 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5bfed592-a73b-4218-bd4a-eab04765a748-config-data\") pod \"glance-default-single-1\" (UID: \"5bfed592-a73b-4218-bd4a-eab04765a748\") " pod="glance-kuttl-tests/glance-default-single-1" Dec 06 03:55:28 crc kubenswrapper[4980]: I1206 03:55:28.793117 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/5bfed592-a73b-4218-bd4a-eab04765a748-etc-iscsi\") pod \"glance-default-single-1\" (UID: \"5bfed592-a73b-4218-bd4a-eab04765a748\") " pod="glance-kuttl-tests/glance-default-single-1" Dec 06 03:55:28 crc kubenswrapper[4980]: I1206 03:55:28.793145 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/5bfed592-a73b-4218-bd4a-eab04765a748-var-locks-brick\") pod \"glance-default-single-1\" (UID: \"5bfed592-a73b-4218-bd4a-eab04765a748\") " pod="glance-kuttl-tests/glance-default-single-1" Dec 06 03:55:28 crc kubenswrapper[4980]: I1206 03:55:28.793178 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-single-1\" (UID: \"5bfed592-a73b-4218-bd4a-eab04765a748\") " pod="glance-kuttl-tests/glance-default-single-1" Dec 06 03:55:28 crc kubenswrapper[4980]: I1206 03:55:28.793201 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/5bfed592-a73b-4218-bd4a-eab04765a748-httpd-run\") pod \"glance-default-single-1\" (UID: \"5bfed592-a73b-4218-bd4a-eab04765a748\") " pod="glance-kuttl-tests/glance-default-single-1" Dec 06 03:55:28 crc kubenswrapper[4980]: I1206 03:55:28.793217 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5bfed592-a73b-4218-bd4a-eab04765a748-logs\") pod \"glance-default-single-1\" (UID: \"5bfed592-a73b-4218-bd4a-eab04765a748\") " pod="glance-kuttl-tests/glance-default-single-1" Dec 06 03:55:28 crc kubenswrapper[4980]: I1206 03:55:28.793236 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/5bfed592-a73b-4218-bd4a-eab04765a748-sys\") pod \"glance-default-single-1\" (UID: \"5bfed592-a73b-4218-bd4a-eab04765a748\") " pod="glance-kuttl-tests/glance-default-single-1" Dec 06 03:55:28 crc kubenswrapper[4980]: I1206 03:55:28.793249 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5bfed592-a73b-4218-bd4a-eab04765a748-scripts\") pod \"glance-default-single-1\" (UID: \"5bfed592-a73b-4218-bd4a-eab04765a748\") " pod="glance-kuttl-tests/glance-default-single-1" Dec 06 03:55:28 crc kubenswrapper[4980]: I1206 03:55:28.793264 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-single-1\" (UID: \"5bfed592-a73b-4218-bd4a-eab04765a748\") " pod="glance-kuttl-tests/glance-default-single-1" Dec 06 03:55:28 crc kubenswrapper[4980]: I1206 03:55:28.793280 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/5bfed592-a73b-4218-bd4a-eab04765a748-dev\") pod \"glance-default-single-1\" (UID: \"5bfed592-a73b-4218-bd4a-eab04765a748\") " pod="glance-kuttl-tests/glance-default-single-1" Dec 06 03:55:28 crc kubenswrapper[4980]: I1206 03:55:28.894601 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/5bfed592-a73b-4218-bd4a-eab04765a748-run\") pod \"glance-default-single-1\" (UID: \"5bfed592-a73b-4218-bd4a-eab04765a748\") " pod="glance-kuttl-tests/glance-default-single-1" Dec 06 03:55:28 crc kubenswrapper[4980]: I1206 03:55:28.894662 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/5bfed592-a73b-4218-bd4a-eab04765a748-etc-nvme\") pod \"glance-default-single-1\" (UID: \"5bfed592-a73b-4218-bd4a-eab04765a748\") " pod="glance-kuttl-tests/glance-default-single-1" Dec 06 03:55:28 crc kubenswrapper[4980]: I1206 03:55:28.894711 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5bfed592-a73b-4218-bd4a-eab04765a748-config-data\") pod \"glance-default-single-1\" (UID: \"5bfed592-a73b-4218-bd4a-eab04765a748\") " pod="glance-kuttl-tests/glance-default-single-1" Dec 06 03:55:28 crc kubenswrapper[4980]: I1206 03:55:28.894734 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/5bfed592-a73b-4218-bd4a-eab04765a748-etc-iscsi\") pod \"glance-default-single-1\" (UID: \"5bfed592-a73b-4218-bd4a-eab04765a748\") " pod="glance-kuttl-tests/glance-default-single-1" Dec 06 03:55:28 crc kubenswrapper[4980]: I1206 03:55:28.894775 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/5bfed592-a73b-4218-bd4a-eab04765a748-var-locks-brick\") pod \"glance-default-single-1\" (UID: \"5bfed592-a73b-4218-bd4a-eab04765a748\") " pod="glance-kuttl-tests/glance-default-single-1" Dec 06 03:55:28 crc kubenswrapper[4980]: I1206 03:55:28.894804 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-single-1\" (UID: \"5bfed592-a73b-4218-bd4a-eab04765a748\") " pod="glance-kuttl-tests/glance-default-single-1" Dec 06 03:55:28 crc kubenswrapper[4980]: I1206 03:55:28.894835 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/5bfed592-a73b-4218-bd4a-eab04765a748-httpd-run\") pod \"glance-default-single-1\" (UID: \"5bfed592-a73b-4218-bd4a-eab04765a748\") " pod="glance-kuttl-tests/glance-default-single-1" Dec 06 03:55:28 crc kubenswrapper[4980]: I1206 03:55:28.894861 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5bfed592-a73b-4218-bd4a-eab04765a748-logs\") pod \"glance-default-single-1\" (UID: \"5bfed592-a73b-4218-bd4a-eab04765a748\") " pod="glance-kuttl-tests/glance-default-single-1" Dec 06 03:55:28 crc kubenswrapper[4980]: I1206 03:55:28.894884 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/5bfed592-a73b-4218-bd4a-eab04765a748-sys\") pod \"glance-default-single-1\" (UID: \"5bfed592-a73b-4218-bd4a-eab04765a748\") " pod="glance-kuttl-tests/glance-default-single-1" Dec 06 03:55:28 crc kubenswrapper[4980]: I1206 03:55:28.894903 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5bfed592-a73b-4218-bd4a-eab04765a748-scripts\") pod \"glance-default-single-1\" (UID: \"5bfed592-a73b-4218-bd4a-eab04765a748\") " pod="glance-kuttl-tests/glance-default-single-1" Dec 06 03:55:28 crc kubenswrapper[4980]: I1206 03:55:28.894938 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-single-1\" (UID: \"5bfed592-a73b-4218-bd4a-eab04765a748\") " pod="glance-kuttl-tests/glance-default-single-1" Dec 06 03:55:28 crc kubenswrapper[4980]: I1206 03:55:28.894963 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/5bfed592-a73b-4218-bd4a-eab04765a748-dev\") pod \"glance-default-single-1\" (UID: \"5bfed592-a73b-4218-bd4a-eab04765a748\") " pod="glance-kuttl-tests/glance-default-single-1" Dec 06 03:55:28 crc kubenswrapper[4980]: I1206 03:55:28.894999 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/5bfed592-a73b-4218-bd4a-eab04765a748-lib-modules\") pod \"glance-default-single-1\" (UID: \"5bfed592-a73b-4218-bd4a-eab04765a748\") " pod="glance-kuttl-tests/glance-default-single-1" Dec 06 03:55:28 crc kubenswrapper[4980]: I1206 03:55:28.895029 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tvc5b\" (UniqueName: \"kubernetes.io/projected/5bfed592-a73b-4218-bd4a-eab04765a748-kube-api-access-tvc5b\") pod \"glance-default-single-1\" (UID: \"5bfed592-a73b-4218-bd4a-eab04765a748\") " pod="glance-kuttl-tests/glance-default-single-1" Dec 06 03:55:28 crc kubenswrapper[4980]: I1206 03:55:28.895398 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/5bfed592-a73b-4218-bd4a-eab04765a748-run\") pod \"glance-default-single-1\" (UID: \"5bfed592-a73b-4218-bd4a-eab04765a748\") " pod="glance-kuttl-tests/glance-default-single-1" Dec 06 03:55:28 crc kubenswrapper[4980]: I1206 03:55:28.895457 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/5bfed592-a73b-4218-bd4a-eab04765a748-etc-nvme\") pod \"glance-default-single-1\" (UID: \"5bfed592-a73b-4218-bd4a-eab04765a748\") " pod="glance-kuttl-tests/glance-default-single-1" Dec 06 03:55:28 crc kubenswrapper[4980]: I1206 03:55:28.896431 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/5bfed592-a73b-4218-bd4a-eab04765a748-dev\") pod \"glance-default-single-1\" (UID: \"5bfed592-a73b-4218-bd4a-eab04765a748\") " pod="glance-kuttl-tests/glance-default-single-1" Dec 06 03:55:28 crc kubenswrapper[4980]: I1206 03:55:28.896481 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/5bfed592-a73b-4218-bd4a-eab04765a748-sys\") pod \"glance-default-single-1\" (UID: \"5bfed592-a73b-4218-bd4a-eab04765a748\") " pod="glance-kuttl-tests/glance-default-single-1" Dec 06 03:55:28 crc kubenswrapper[4980]: I1206 03:55:28.896530 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/5bfed592-a73b-4218-bd4a-eab04765a748-etc-iscsi\") pod \"glance-default-single-1\" (UID: \"5bfed592-a73b-4218-bd4a-eab04765a748\") " pod="glance-kuttl-tests/glance-default-single-1" Dec 06 03:55:28 crc kubenswrapper[4980]: I1206 03:55:28.896533 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/5bfed592-a73b-4218-bd4a-eab04765a748-var-locks-brick\") pod \"glance-default-single-1\" (UID: \"5bfed592-a73b-4218-bd4a-eab04765a748\") " pod="glance-kuttl-tests/glance-default-single-1" Dec 06 03:55:28 crc kubenswrapper[4980]: I1206 03:55:28.896530 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/5bfed592-a73b-4218-bd4a-eab04765a748-lib-modules\") pod \"glance-default-single-1\" (UID: \"5bfed592-a73b-4218-bd4a-eab04765a748\") " pod="glance-kuttl-tests/glance-default-single-1" Dec 06 03:55:28 crc kubenswrapper[4980]: I1206 03:55:28.896818 4980 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-single-1\" (UID: \"5bfed592-a73b-4218-bd4a-eab04765a748\") device mount path \"/mnt/openstack/pv08\"" pod="glance-kuttl-tests/glance-default-single-1" Dec 06 03:55:28 crc kubenswrapper[4980]: I1206 03:55:28.896975 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/5bfed592-a73b-4218-bd4a-eab04765a748-httpd-run\") pod \"glance-default-single-1\" (UID: \"5bfed592-a73b-4218-bd4a-eab04765a748\") " pod="glance-kuttl-tests/glance-default-single-1" Dec 06 03:55:28 crc kubenswrapper[4980]: I1206 03:55:28.897016 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5bfed592-a73b-4218-bd4a-eab04765a748-logs\") pod \"glance-default-single-1\" (UID: \"5bfed592-a73b-4218-bd4a-eab04765a748\") " pod="glance-kuttl-tests/glance-default-single-1" Dec 06 03:55:28 crc kubenswrapper[4980]: I1206 03:55:28.897228 4980 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-single-1\" (UID: \"5bfed592-a73b-4218-bd4a-eab04765a748\") device mount path \"/mnt/openstack/pv07\"" pod="glance-kuttl-tests/glance-default-single-1" Dec 06 03:55:28 crc kubenswrapper[4980]: I1206 03:55:28.902163 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5bfed592-a73b-4218-bd4a-eab04765a748-scripts\") pod \"glance-default-single-1\" (UID: \"5bfed592-a73b-4218-bd4a-eab04765a748\") " pod="glance-kuttl-tests/glance-default-single-1" Dec 06 03:55:28 crc kubenswrapper[4980]: I1206 03:55:28.903413 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5bfed592-a73b-4218-bd4a-eab04765a748-config-data\") pod \"glance-default-single-1\" (UID: \"5bfed592-a73b-4218-bd4a-eab04765a748\") " pod="glance-kuttl-tests/glance-default-single-1" Dec 06 03:55:28 crc kubenswrapper[4980]: I1206 03:55:28.916014 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tvc5b\" (UniqueName: \"kubernetes.io/projected/5bfed592-a73b-4218-bd4a-eab04765a748-kube-api-access-tvc5b\") pod \"glance-default-single-1\" (UID: \"5bfed592-a73b-4218-bd4a-eab04765a748\") " pod="glance-kuttl-tests/glance-default-single-1" Dec 06 03:55:28 crc kubenswrapper[4980]: I1206 03:55:28.921708 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-single-1\" (UID: \"5bfed592-a73b-4218-bd4a-eab04765a748\") " pod="glance-kuttl-tests/glance-default-single-1" Dec 06 03:55:28 crc kubenswrapper[4980]: I1206 03:55:28.924697 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-single-1\" (UID: \"5bfed592-a73b-4218-bd4a-eab04765a748\") " pod="glance-kuttl-tests/glance-default-single-1" Dec 06 03:55:29 crc kubenswrapper[4980]: I1206 03:55:29.193682 4980 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="006f4ba2-a479-400f-99c6-468450cbe52f" path="/var/lib/kubelet/pods/006f4ba2-a479-400f-99c6-468450cbe52f/volumes" Dec 06 03:55:29 crc kubenswrapper[4980]: I1206 03:55:29.219142 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-1" Dec 06 03:55:29 crc kubenswrapper[4980]: I1206 03:55:29.680575 4980 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-single-1"] Dec 06 03:55:29 crc kubenswrapper[4980]: W1206 03:55:29.686422 4980 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5bfed592_a73b_4218_bd4a_eab04765a748.slice/crio-96c7c655066bf28e24a9711be703ae81dfa33e701fdf7f5874d9586601dc3bab WatchSource:0}: Error finding container 96c7c655066bf28e24a9711be703ae81dfa33e701fdf7f5874d9586601dc3bab: Status 404 returned error can't find the container with id 96c7c655066bf28e24a9711be703ae81dfa33e701fdf7f5874d9586601dc3bab Dec 06 03:55:30 crc kubenswrapper[4980]: I1206 03:55:30.468395 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-1" event={"ID":"5bfed592-a73b-4218-bd4a-eab04765a748","Type":"ContainerStarted","Data":"ccfd68de53a61d766d5202a884a95ac78a14bdcda5e296b2a9ece932bde45d0a"} Dec 06 03:55:30 crc kubenswrapper[4980]: I1206 03:55:30.469106 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-1" event={"ID":"5bfed592-a73b-4218-bd4a-eab04765a748","Type":"ContainerStarted","Data":"b89a93f06cb4564f65b336ba76c926fcf3f7cf40cc9bf58d7af24dd958e2a086"} Dec 06 03:55:30 crc kubenswrapper[4980]: I1206 03:55:30.469152 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-1" event={"ID":"5bfed592-a73b-4218-bd4a-eab04765a748","Type":"ContainerStarted","Data":"96c7c655066bf28e24a9711be703ae81dfa33e701fdf7f5874d9586601dc3bab"} Dec 06 03:55:30 crc kubenswrapper[4980]: I1206 03:55:30.493240 4980 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/glance-default-single-1" podStartSLOduration=2.493216844 podStartE2EDuration="2.493216844s" podCreationTimestamp="2025-12-06 03:55:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 03:55:30.488001744 +0000 UTC m=+1289.732753015" watchObservedRunningTime="2025-12-06 03:55:30.493216844 +0000 UTC m=+1289.737968135" Dec 06 03:55:37 crc kubenswrapper[4980]: I1206 03:55:37.425273 4980 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-single-0" Dec 06 03:55:37 crc kubenswrapper[4980]: I1206 03:55:37.426188 4980 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-single-0" Dec 06 03:55:37 crc kubenswrapper[4980]: I1206 03:55:37.456728 4980 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-single-0" Dec 06 03:55:37 crc kubenswrapper[4980]: I1206 03:55:37.474583 4980 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-single-0" Dec 06 03:55:37 crc kubenswrapper[4980]: I1206 03:55:37.530437 4980 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-single-0" Dec 06 03:55:37 crc kubenswrapper[4980]: I1206 03:55:37.530479 4980 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-single-0" Dec 06 03:55:39 crc kubenswrapper[4980]: I1206 03:55:39.219554 4980 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-single-1" Dec 06 03:55:39 crc kubenswrapper[4980]: I1206 03:55:39.219901 4980 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-single-1" Dec 06 03:55:39 crc kubenswrapper[4980]: I1206 03:55:39.243863 4980 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-single-1" Dec 06 03:55:39 crc kubenswrapper[4980]: I1206 03:55:39.271525 4980 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-single-1" Dec 06 03:55:39 crc kubenswrapper[4980]: I1206 03:55:39.573953 4980 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-single-1" Dec 06 03:55:39 crc kubenswrapper[4980]: I1206 03:55:39.574014 4980 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-single-1" Dec 06 03:55:39 crc kubenswrapper[4980]: I1206 03:55:39.664810 4980 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-single-0" Dec 06 03:55:39 crc kubenswrapper[4980]: I1206 03:55:39.665000 4980 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 06 03:55:39 crc kubenswrapper[4980]: I1206 03:55:39.670055 4980 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-single-0" Dec 06 03:55:41 crc kubenswrapper[4980]: I1206 03:55:41.759610 4980 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-single-1" Dec 06 03:55:41 crc kubenswrapper[4980]: I1206 03:55:41.760590 4980 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-single-1" Dec 06 03:55:41 crc kubenswrapper[4980]: I1206 03:55:41.834753 4980 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Dec 06 03:55:41 crc kubenswrapper[4980]: I1206 03:55:41.835086 4980 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-single-0" podUID="de2e2787-7471-41a3-9bd7-e591410f3f42" containerName="glance-log" containerID="cri-o://a98388d0fc55b4344bba2f679d8609c5a5c9f6019d03a5ae0d2772d41450746f" gracePeriod=30 Dec 06 03:55:41 crc kubenswrapper[4980]: I1206 03:55:41.835262 4980 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-single-0" podUID="de2e2787-7471-41a3-9bd7-e591410f3f42" containerName="glance-httpd" containerID="cri-o://1432679ad8686abee0ae98476311d7a88fc012bdb4aeb04483111362fc779aa6" gracePeriod=30 Dec 06 03:55:42 crc kubenswrapper[4980]: I1206 03:55:42.621070 4980 generic.go:334] "Generic (PLEG): container finished" podID="de2e2787-7471-41a3-9bd7-e591410f3f42" containerID="a98388d0fc55b4344bba2f679d8609c5a5c9f6019d03a5ae0d2772d41450746f" exitCode=143 Dec 06 03:55:42 crc kubenswrapper[4980]: I1206 03:55:42.621186 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"de2e2787-7471-41a3-9bd7-e591410f3f42","Type":"ContainerDied","Data":"a98388d0fc55b4344bba2f679d8609c5a5c9f6019d03a5ae0d2772d41450746f"} Dec 06 03:55:45 crc kubenswrapper[4980]: I1206 03:55:45.499039 4980 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-0" Dec 06 03:55:45 crc kubenswrapper[4980]: I1206 03:55:45.647182 4980 generic.go:334] "Generic (PLEG): container finished" podID="de2e2787-7471-41a3-9bd7-e591410f3f42" containerID="1432679ad8686abee0ae98476311d7a88fc012bdb4aeb04483111362fc779aa6" exitCode=0 Dec 06 03:55:45 crc kubenswrapper[4980]: I1206 03:55:45.647231 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"de2e2787-7471-41a3-9bd7-e591410f3f42","Type":"ContainerDied","Data":"1432679ad8686abee0ae98476311d7a88fc012bdb4aeb04483111362fc779aa6"} Dec 06 03:55:45 crc kubenswrapper[4980]: I1206 03:55:45.647272 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"de2e2787-7471-41a3-9bd7-e591410f3f42","Type":"ContainerDied","Data":"74f970886f6dbee3a5ce01e19cc473fac456170f227e1a7c881088af281cabd7"} Dec 06 03:55:45 crc kubenswrapper[4980]: I1206 03:55:45.647289 4980 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-0" Dec 06 03:55:45 crc kubenswrapper[4980]: I1206 03:55:45.647383 4980 scope.go:117] "RemoveContainer" containerID="1432679ad8686abee0ae98476311d7a88fc012bdb4aeb04483111362fc779aa6" Dec 06 03:55:45 crc kubenswrapper[4980]: I1206 03:55:45.672245 4980 scope.go:117] "RemoveContainer" containerID="a98388d0fc55b4344bba2f679d8609c5a5c9f6019d03a5ae0d2772d41450746f" Dec 06 03:55:45 crc kubenswrapper[4980]: I1206 03:55:45.690444 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/de2e2787-7471-41a3-9bd7-e591410f3f42-sys\") pod \"de2e2787-7471-41a3-9bd7-e591410f3f42\" (UID: \"de2e2787-7471-41a3-9bd7-e591410f3f42\") " Dec 06 03:55:45 crc kubenswrapper[4980]: I1206 03:55:45.690545 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/de2e2787-7471-41a3-9bd7-e591410f3f42-etc-iscsi\") pod \"de2e2787-7471-41a3-9bd7-e591410f3f42\" (UID: \"de2e2787-7471-41a3-9bd7-e591410f3f42\") " Dec 06 03:55:45 crc kubenswrapper[4980]: I1206 03:55:45.690587 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/de2e2787-7471-41a3-9bd7-e591410f3f42-etc-nvme\") pod \"de2e2787-7471-41a3-9bd7-e591410f3f42\" (UID: \"de2e2787-7471-41a3-9bd7-e591410f3f42\") " Dec 06 03:55:45 crc kubenswrapper[4980]: I1206 03:55:45.690642 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/de2e2787-7471-41a3-9bd7-e591410f3f42-sys" (OuterVolumeSpecName: "sys") pod "de2e2787-7471-41a3-9bd7-e591410f3f42" (UID: "de2e2787-7471-41a3-9bd7-e591410f3f42"). InnerVolumeSpecName "sys". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 03:55:45 crc kubenswrapper[4980]: I1206 03:55:45.690668 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/de2e2787-7471-41a3-9bd7-e591410f3f42-httpd-run\") pod \"de2e2787-7471-41a3-9bd7-e591410f3f42\" (UID: \"de2e2787-7471-41a3-9bd7-e591410f3f42\") " Dec 06 03:55:45 crc kubenswrapper[4980]: I1206 03:55:45.690762 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/de2e2787-7471-41a3-9bd7-e591410f3f42-lib-modules\") pod \"de2e2787-7471-41a3-9bd7-e591410f3f42\" (UID: \"de2e2787-7471-41a3-9bd7-e591410f3f42\") " Dec 06 03:55:45 crc kubenswrapper[4980]: I1206 03:55:45.690802 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/de2e2787-7471-41a3-9bd7-e591410f3f42-run\") pod \"de2e2787-7471-41a3-9bd7-e591410f3f42\" (UID: \"de2e2787-7471-41a3-9bd7-e591410f3f42\") " Dec 06 03:55:45 crc kubenswrapper[4980]: I1206 03:55:45.690703 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/de2e2787-7471-41a3-9bd7-e591410f3f42-etc-iscsi" (OuterVolumeSpecName: "etc-iscsi") pod "de2e2787-7471-41a3-9bd7-e591410f3f42" (UID: "de2e2787-7471-41a3-9bd7-e591410f3f42"). InnerVolumeSpecName "etc-iscsi". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 03:55:45 crc kubenswrapper[4980]: I1206 03:55:45.690729 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/de2e2787-7471-41a3-9bd7-e591410f3f42-etc-nvme" (OuterVolumeSpecName: "etc-nvme") pod "de2e2787-7471-41a3-9bd7-e591410f3f42" (UID: "de2e2787-7471-41a3-9bd7-e591410f3f42"). InnerVolumeSpecName "etc-nvme". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 03:55:45 crc kubenswrapper[4980]: I1206 03:55:45.690861 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-99bgs\" (UniqueName: \"kubernetes.io/projected/de2e2787-7471-41a3-9bd7-e591410f3f42-kube-api-access-99bgs\") pod \"de2e2787-7471-41a3-9bd7-e591410f3f42\" (UID: \"de2e2787-7471-41a3-9bd7-e591410f3f42\") " Dec 06 03:55:45 crc kubenswrapper[4980]: I1206 03:55:45.690871 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/de2e2787-7471-41a3-9bd7-e591410f3f42-lib-modules" (OuterVolumeSpecName: "lib-modules") pod "de2e2787-7471-41a3-9bd7-e591410f3f42" (UID: "de2e2787-7471-41a3-9bd7-e591410f3f42"). InnerVolumeSpecName "lib-modules". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 03:55:45 crc kubenswrapper[4980]: I1206 03:55:45.690894 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/de2e2787-7471-41a3-9bd7-e591410f3f42-config-data\") pod \"de2e2787-7471-41a3-9bd7-e591410f3f42\" (UID: \"de2e2787-7471-41a3-9bd7-e591410f3f42\") " Dec 06 03:55:45 crc kubenswrapper[4980]: I1206 03:55:45.690935 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/de2e2787-7471-41a3-9bd7-e591410f3f42-scripts\") pod \"de2e2787-7471-41a3-9bd7-e591410f3f42\" (UID: \"de2e2787-7471-41a3-9bd7-e591410f3f42\") " Dec 06 03:55:45 crc kubenswrapper[4980]: I1206 03:55:45.691000 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"de2e2787-7471-41a3-9bd7-e591410f3f42\" (UID: \"de2e2787-7471-41a3-9bd7-e591410f3f42\") " Dec 06 03:55:45 crc kubenswrapper[4980]: I1206 03:55:45.691044 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/de2e2787-7471-41a3-9bd7-e591410f3f42-dev\") pod \"de2e2787-7471-41a3-9bd7-e591410f3f42\" (UID: \"de2e2787-7471-41a3-9bd7-e591410f3f42\") " Dec 06 03:55:45 crc kubenswrapper[4980]: I1206 03:55:45.691076 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/de2e2787-7471-41a3-9bd7-e591410f3f42-var-locks-brick\") pod \"de2e2787-7471-41a3-9bd7-e591410f3f42\" (UID: \"de2e2787-7471-41a3-9bd7-e591410f3f42\") " Dec 06 03:55:45 crc kubenswrapper[4980]: I1206 03:55:45.691144 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/de2e2787-7471-41a3-9bd7-e591410f3f42-logs\") pod \"de2e2787-7471-41a3-9bd7-e591410f3f42\" (UID: \"de2e2787-7471-41a3-9bd7-e591410f3f42\") " Dec 06 03:55:45 crc kubenswrapper[4980]: I1206 03:55:45.691178 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance-cache\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"de2e2787-7471-41a3-9bd7-e591410f3f42\" (UID: \"de2e2787-7471-41a3-9bd7-e591410f3f42\") " Dec 06 03:55:45 crc kubenswrapper[4980]: I1206 03:55:45.691807 4980 reconciler_common.go:293] "Volume detached for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/de2e2787-7471-41a3-9bd7-e591410f3f42-etc-iscsi\") on node \"crc\" DevicePath \"\"" Dec 06 03:55:45 crc kubenswrapper[4980]: I1206 03:55:45.691839 4980 reconciler_common.go:293] "Volume detached for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/de2e2787-7471-41a3-9bd7-e591410f3f42-etc-nvme\") on node \"crc\" DevicePath \"\"" Dec 06 03:55:45 crc kubenswrapper[4980]: I1206 03:55:45.691856 4980 reconciler_common.go:293] "Volume detached for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/de2e2787-7471-41a3-9bd7-e591410f3f42-lib-modules\") on node \"crc\" DevicePath \"\"" Dec 06 03:55:45 crc kubenswrapper[4980]: I1206 03:55:45.691875 4980 reconciler_common.go:293] "Volume detached for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/de2e2787-7471-41a3-9bd7-e591410f3f42-sys\") on node \"crc\" DevicePath \"\"" Dec 06 03:55:45 crc kubenswrapper[4980]: I1206 03:55:45.690899 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/de2e2787-7471-41a3-9bd7-e591410f3f42-run" (OuterVolumeSpecName: "run") pod "de2e2787-7471-41a3-9bd7-e591410f3f42" (UID: "de2e2787-7471-41a3-9bd7-e591410f3f42"). InnerVolumeSpecName "run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 03:55:45 crc kubenswrapper[4980]: I1206 03:55:45.691081 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/de2e2787-7471-41a3-9bd7-e591410f3f42-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "de2e2787-7471-41a3-9bd7-e591410f3f42" (UID: "de2e2787-7471-41a3-9bd7-e591410f3f42"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 03:55:45 crc kubenswrapper[4980]: I1206 03:55:45.691792 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/de2e2787-7471-41a3-9bd7-e591410f3f42-dev" (OuterVolumeSpecName: "dev") pod "de2e2787-7471-41a3-9bd7-e591410f3f42" (UID: "de2e2787-7471-41a3-9bd7-e591410f3f42"). InnerVolumeSpecName "dev". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 03:55:45 crc kubenswrapper[4980]: I1206 03:55:45.691935 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/de2e2787-7471-41a3-9bd7-e591410f3f42-var-locks-brick" (OuterVolumeSpecName: "var-locks-brick") pod "de2e2787-7471-41a3-9bd7-e591410f3f42" (UID: "de2e2787-7471-41a3-9bd7-e591410f3f42"). InnerVolumeSpecName "var-locks-brick". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 03:55:45 crc kubenswrapper[4980]: I1206 03:55:45.692376 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/de2e2787-7471-41a3-9bd7-e591410f3f42-logs" (OuterVolumeSpecName: "logs") pod "de2e2787-7471-41a3-9bd7-e591410f3f42" (UID: "de2e2787-7471-41a3-9bd7-e591410f3f42"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 03:55:45 crc kubenswrapper[4980]: I1206 03:55:45.694073 4980 scope.go:117] "RemoveContainer" containerID="1432679ad8686abee0ae98476311d7a88fc012bdb4aeb04483111362fc779aa6" Dec 06 03:55:45 crc kubenswrapper[4980]: E1206 03:55:45.694645 4980 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1432679ad8686abee0ae98476311d7a88fc012bdb4aeb04483111362fc779aa6\": container with ID starting with 1432679ad8686abee0ae98476311d7a88fc012bdb4aeb04483111362fc779aa6 not found: ID does not exist" containerID="1432679ad8686abee0ae98476311d7a88fc012bdb4aeb04483111362fc779aa6" Dec 06 03:55:45 crc kubenswrapper[4980]: I1206 03:55:45.694694 4980 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1432679ad8686abee0ae98476311d7a88fc012bdb4aeb04483111362fc779aa6"} err="failed to get container status \"1432679ad8686abee0ae98476311d7a88fc012bdb4aeb04483111362fc779aa6\": rpc error: code = NotFound desc = could not find container \"1432679ad8686abee0ae98476311d7a88fc012bdb4aeb04483111362fc779aa6\": container with ID starting with 1432679ad8686abee0ae98476311d7a88fc012bdb4aeb04483111362fc779aa6 not found: ID does not exist" Dec 06 03:55:45 crc kubenswrapper[4980]: I1206 03:55:45.694728 4980 scope.go:117] "RemoveContainer" containerID="a98388d0fc55b4344bba2f679d8609c5a5c9f6019d03a5ae0d2772d41450746f" Dec 06 03:55:45 crc kubenswrapper[4980]: E1206 03:55:45.694998 4980 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a98388d0fc55b4344bba2f679d8609c5a5c9f6019d03a5ae0d2772d41450746f\": container with ID starting with a98388d0fc55b4344bba2f679d8609c5a5c9f6019d03a5ae0d2772d41450746f not found: ID does not exist" containerID="a98388d0fc55b4344bba2f679d8609c5a5c9f6019d03a5ae0d2772d41450746f" Dec 06 03:55:45 crc kubenswrapper[4980]: I1206 03:55:45.695230 4980 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a98388d0fc55b4344bba2f679d8609c5a5c9f6019d03a5ae0d2772d41450746f"} err="failed to get container status \"a98388d0fc55b4344bba2f679d8609c5a5c9f6019d03a5ae0d2772d41450746f\": rpc error: code = NotFound desc = could not find container \"a98388d0fc55b4344bba2f679d8609c5a5c9f6019d03a5ae0d2772d41450746f\": container with ID starting with a98388d0fc55b4344bba2f679d8609c5a5c9f6019d03a5ae0d2772d41450746f not found: ID does not exist" Dec 06 03:55:45 crc kubenswrapper[4980]: I1206 03:55:45.696488 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage09-crc" (OuterVolumeSpecName: "glance") pod "de2e2787-7471-41a3-9bd7-e591410f3f42" (UID: "de2e2787-7471-41a3-9bd7-e591410f3f42"). InnerVolumeSpecName "local-storage09-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 06 03:55:45 crc kubenswrapper[4980]: I1206 03:55:45.699733 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/de2e2787-7471-41a3-9bd7-e591410f3f42-kube-api-access-99bgs" (OuterVolumeSpecName: "kube-api-access-99bgs") pod "de2e2787-7471-41a3-9bd7-e591410f3f42" (UID: "de2e2787-7471-41a3-9bd7-e591410f3f42"). InnerVolumeSpecName "kube-api-access-99bgs". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 03:55:45 crc kubenswrapper[4980]: I1206 03:55:45.701151 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage05-crc" (OuterVolumeSpecName: "glance-cache") pod "de2e2787-7471-41a3-9bd7-e591410f3f42" (UID: "de2e2787-7471-41a3-9bd7-e591410f3f42"). InnerVolumeSpecName "local-storage05-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 06 03:55:45 crc kubenswrapper[4980]: I1206 03:55:45.701988 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/de2e2787-7471-41a3-9bd7-e591410f3f42-scripts" (OuterVolumeSpecName: "scripts") pod "de2e2787-7471-41a3-9bd7-e591410f3f42" (UID: "de2e2787-7471-41a3-9bd7-e591410f3f42"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 03:55:45 crc kubenswrapper[4980]: I1206 03:55:45.752897 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/de2e2787-7471-41a3-9bd7-e591410f3f42-config-data" (OuterVolumeSpecName: "config-data") pod "de2e2787-7471-41a3-9bd7-e591410f3f42" (UID: "de2e2787-7471-41a3-9bd7-e591410f3f42"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 03:55:45 crc kubenswrapper[4980]: I1206 03:55:45.793786 4980 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/de2e2787-7471-41a3-9bd7-e591410f3f42-httpd-run\") on node \"crc\" DevicePath \"\"" Dec 06 03:55:45 crc kubenswrapper[4980]: I1206 03:55:45.793820 4980 reconciler_common.go:293] "Volume detached for volume \"run\" (UniqueName: \"kubernetes.io/host-path/de2e2787-7471-41a3-9bd7-e591410f3f42-run\") on node \"crc\" DevicePath \"\"" Dec 06 03:55:45 crc kubenswrapper[4980]: I1206 03:55:45.793829 4980 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-99bgs\" (UniqueName: \"kubernetes.io/projected/de2e2787-7471-41a3-9bd7-e591410f3f42-kube-api-access-99bgs\") on node \"crc\" DevicePath \"\"" Dec 06 03:55:45 crc kubenswrapper[4980]: I1206 03:55:45.793840 4980 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/de2e2787-7471-41a3-9bd7-e591410f3f42-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 03:55:45 crc kubenswrapper[4980]: I1206 03:55:45.793848 4980 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/de2e2787-7471-41a3-9bd7-e591410f3f42-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 03:55:45 crc kubenswrapper[4980]: I1206 03:55:45.793890 4980 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") on node \"crc\" " Dec 06 03:55:45 crc kubenswrapper[4980]: I1206 03:55:45.793900 4980 reconciler_common.go:293] "Volume detached for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/de2e2787-7471-41a3-9bd7-e591410f3f42-dev\") on node \"crc\" DevicePath \"\"" Dec 06 03:55:45 crc kubenswrapper[4980]: I1206 03:55:45.793908 4980 reconciler_common.go:293] "Volume detached for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/de2e2787-7471-41a3-9bd7-e591410f3f42-var-locks-brick\") on node \"crc\" DevicePath \"\"" Dec 06 03:55:45 crc kubenswrapper[4980]: I1206 03:55:45.793917 4980 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/de2e2787-7471-41a3-9bd7-e591410f3f42-logs\") on node \"crc\" DevicePath \"\"" Dec 06 03:55:45 crc kubenswrapper[4980]: I1206 03:55:45.793932 4980 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") on node \"crc\" " Dec 06 03:55:45 crc kubenswrapper[4980]: I1206 03:55:45.816055 4980 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage09-crc" (UniqueName: "kubernetes.io/local-volume/local-storage09-crc") on node "crc" Dec 06 03:55:45 crc kubenswrapper[4980]: I1206 03:55:45.822737 4980 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage05-crc" (UniqueName: "kubernetes.io/local-volume/local-storage05-crc") on node "crc" Dec 06 03:55:45 crc kubenswrapper[4980]: I1206 03:55:45.895451 4980 reconciler_common.go:293] "Volume detached for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") on node \"crc\" DevicePath \"\"" Dec 06 03:55:45 crc kubenswrapper[4980]: I1206 03:55:45.895487 4980 reconciler_common.go:293] "Volume detached for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") on node \"crc\" DevicePath \"\"" Dec 06 03:55:45 crc kubenswrapper[4980]: I1206 03:55:45.977533 4980 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Dec 06 03:55:45 crc kubenswrapper[4980]: I1206 03:55:45.996834 4980 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Dec 06 03:55:46 crc kubenswrapper[4980]: I1206 03:55:46.043161 4980 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Dec 06 03:55:46 crc kubenswrapper[4980]: E1206 03:55:46.043556 4980 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="de2e2787-7471-41a3-9bd7-e591410f3f42" containerName="glance-httpd" Dec 06 03:55:46 crc kubenswrapper[4980]: I1206 03:55:46.043578 4980 state_mem.go:107] "Deleted CPUSet assignment" podUID="de2e2787-7471-41a3-9bd7-e591410f3f42" containerName="glance-httpd" Dec 06 03:55:46 crc kubenswrapper[4980]: E1206 03:55:46.043605 4980 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="de2e2787-7471-41a3-9bd7-e591410f3f42" containerName="glance-log" Dec 06 03:55:46 crc kubenswrapper[4980]: I1206 03:55:46.043614 4980 state_mem.go:107] "Deleted CPUSet assignment" podUID="de2e2787-7471-41a3-9bd7-e591410f3f42" containerName="glance-log" Dec 06 03:55:46 crc kubenswrapper[4980]: I1206 03:55:46.043815 4980 memory_manager.go:354] "RemoveStaleState removing state" podUID="de2e2787-7471-41a3-9bd7-e591410f3f42" containerName="glance-log" Dec 06 03:55:46 crc kubenswrapper[4980]: I1206 03:55:46.043862 4980 memory_manager.go:354] "RemoveStaleState removing state" podUID="de2e2787-7471-41a3-9bd7-e591410f3f42" containerName="glance-httpd" Dec 06 03:55:46 crc kubenswrapper[4980]: I1206 03:55:46.044892 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-0" Dec 06 03:55:46 crc kubenswrapper[4980]: I1206 03:55:46.072532 4980 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Dec 06 03:55:46 crc kubenswrapper[4980]: I1206 03:55:46.198858 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7rpnn\" (UniqueName: \"kubernetes.io/projected/ce3481e1-ecb6-4d13-9258-656e9ac4559d-kube-api-access-7rpnn\") pod \"glance-default-single-0\" (UID: \"ce3481e1-ecb6-4d13-9258-656e9ac4559d\") " pod="glance-kuttl-tests/glance-default-single-0" Dec 06 03:55:46 crc kubenswrapper[4980]: I1206 03:55:46.198913 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ce3481e1-ecb6-4d13-9258-656e9ac4559d-scripts\") pod \"glance-default-single-0\" (UID: \"ce3481e1-ecb6-4d13-9258-656e9ac4559d\") " pod="glance-kuttl-tests/glance-default-single-0" Dec 06 03:55:46 crc kubenswrapper[4980]: I1206 03:55:46.198937 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/ce3481e1-ecb6-4d13-9258-656e9ac4559d-var-locks-brick\") pod \"glance-default-single-0\" (UID: \"ce3481e1-ecb6-4d13-9258-656e9ac4559d\") " pod="glance-kuttl-tests/glance-default-single-0" Dec 06 03:55:46 crc kubenswrapper[4980]: I1206 03:55:46.198957 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/ce3481e1-ecb6-4d13-9258-656e9ac4559d-run\") pod \"glance-default-single-0\" (UID: \"ce3481e1-ecb6-4d13-9258-656e9ac4559d\") " pod="glance-kuttl-tests/glance-default-single-0" Dec 06 03:55:46 crc kubenswrapper[4980]: I1206 03:55:46.198973 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/ce3481e1-ecb6-4d13-9258-656e9ac4559d-etc-iscsi\") pod \"glance-default-single-0\" (UID: \"ce3481e1-ecb6-4d13-9258-656e9ac4559d\") " pod="glance-kuttl-tests/glance-default-single-0" Dec 06 03:55:46 crc kubenswrapper[4980]: I1206 03:55:46.198988 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/ce3481e1-ecb6-4d13-9258-656e9ac4559d-httpd-run\") pod \"glance-default-single-0\" (UID: \"ce3481e1-ecb6-4d13-9258-656e9ac4559d\") " pod="glance-kuttl-tests/glance-default-single-0" Dec 06 03:55:46 crc kubenswrapper[4980]: I1206 03:55:46.199061 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ce3481e1-ecb6-4d13-9258-656e9ac4559d-config-data\") pod \"glance-default-single-0\" (UID: \"ce3481e1-ecb6-4d13-9258-656e9ac4559d\") " pod="glance-kuttl-tests/glance-default-single-0" Dec 06 03:55:46 crc kubenswrapper[4980]: I1206 03:55:46.199119 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/ce3481e1-ecb6-4d13-9258-656e9ac4559d-lib-modules\") pod \"glance-default-single-0\" (UID: \"ce3481e1-ecb6-4d13-9258-656e9ac4559d\") " pod="glance-kuttl-tests/glance-default-single-0" Dec 06 03:55:46 crc kubenswrapper[4980]: I1206 03:55:46.199143 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/ce3481e1-ecb6-4d13-9258-656e9ac4559d-dev\") pod \"glance-default-single-0\" (UID: \"ce3481e1-ecb6-4d13-9258-656e9ac4559d\") " pod="glance-kuttl-tests/glance-default-single-0" Dec 06 03:55:46 crc kubenswrapper[4980]: I1206 03:55:46.199159 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/ce3481e1-ecb6-4d13-9258-656e9ac4559d-sys\") pod \"glance-default-single-0\" (UID: \"ce3481e1-ecb6-4d13-9258-656e9ac4559d\") " pod="glance-kuttl-tests/glance-default-single-0" Dec 06 03:55:46 crc kubenswrapper[4980]: I1206 03:55:46.199177 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-single-0\" (UID: \"ce3481e1-ecb6-4d13-9258-656e9ac4559d\") " pod="glance-kuttl-tests/glance-default-single-0" Dec 06 03:55:46 crc kubenswrapper[4980]: I1206 03:55:46.199214 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/ce3481e1-ecb6-4d13-9258-656e9ac4559d-etc-nvme\") pod \"glance-default-single-0\" (UID: \"ce3481e1-ecb6-4d13-9258-656e9ac4559d\") " pod="glance-kuttl-tests/glance-default-single-0" Dec 06 03:55:46 crc kubenswrapper[4980]: I1206 03:55:46.199367 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ce3481e1-ecb6-4d13-9258-656e9ac4559d-logs\") pod \"glance-default-single-0\" (UID: \"ce3481e1-ecb6-4d13-9258-656e9ac4559d\") " pod="glance-kuttl-tests/glance-default-single-0" Dec 06 03:55:46 crc kubenswrapper[4980]: I1206 03:55:46.199448 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-single-0\" (UID: \"ce3481e1-ecb6-4d13-9258-656e9ac4559d\") " pod="glance-kuttl-tests/glance-default-single-0" Dec 06 03:55:46 crc kubenswrapper[4980]: I1206 03:55:46.300540 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/ce3481e1-ecb6-4d13-9258-656e9ac4559d-etc-nvme\") pod \"glance-default-single-0\" (UID: \"ce3481e1-ecb6-4d13-9258-656e9ac4559d\") " pod="glance-kuttl-tests/glance-default-single-0" Dec 06 03:55:46 crc kubenswrapper[4980]: I1206 03:55:46.300586 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/ce3481e1-ecb6-4d13-9258-656e9ac4559d-etc-nvme\") pod \"glance-default-single-0\" (UID: \"ce3481e1-ecb6-4d13-9258-656e9ac4559d\") " pod="glance-kuttl-tests/glance-default-single-0" Dec 06 03:55:46 crc kubenswrapper[4980]: I1206 03:55:46.300610 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ce3481e1-ecb6-4d13-9258-656e9ac4559d-logs\") pod \"glance-default-single-0\" (UID: \"ce3481e1-ecb6-4d13-9258-656e9ac4559d\") " pod="glance-kuttl-tests/glance-default-single-0" Dec 06 03:55:46 crc kubenswrapper[4980]: I1206 03:55:46.300663 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-single-0\" (UID: \"ce3481e1-ecb6-4d13-9258-656e9ac4559d\") " pod="glance-kuttl-tests/glance-default-single-0" Dec 06 03:55:46 crc kubenswrapper[4980]: I1206 03:55:46.300720 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7rpnn\" (UniqueName: \"kubernetes.io/projected/ce3481e1-ecb6-4d13-9258-656e9ac4559d-kube-api-access-7rpnn\") pod \"glance-default-single-0\" (UID: \"ce3481e1-ecb6-4d13-9258-656e9ac4559d\") " pod="glance-kuttl-tests/glance-default-single-0" Dec 06 03:55:46 crc kubenswrapper[4980]: I1206 03:55:46.300777 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/ce3481e1-ecb6-4d13-9258-656e9ac4559d-var-locks-brick\") pod \"glance-default-single-0\" (UID: \"ce3481e1-ecb6-4d13-9258-656e9ac4559d\") " pod="glance-kuttl-tests/glance-default-single-0" Dec 06 03:55:46 crc kubenswrapper[4980]: I1206 03:55:46.300805 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ce3481e1-ecb6-4d13-9258-656e9ac4559d-scripts\") pod \"glance-default-single-0\" (UID: \"ce3481e1-ecb6-4d13-9258-656e9ac4559d\") " pod="glance-kuttl-tests/glance-default-single-0" Dec 06 03:55:46 crc kubenswrapper[4980]: I1206 03:55:46.300831 4980 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-single-0\" (UID: \"ce3481e1-ecb6-4d13-9258-656e9ac4559d\") device mount path \"/mnt/openstack/pv09\"" pod="glance-kuttl-tests/glance-default-single-0" Dec 06 03:55:46 crc kubenswrapper[4980]: I1206 03:55:46.300971 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/ce3481e1-ecb6-4d13-9258-656e9ac4559d-var-locks-brick\") pod \"glance-default-single-0\" (UID: \"ce3481e1-ecb6-4d13-9258-656e9ac4559d\") " pod="glance-kuttl-tests/glance-default-single-0" Dec 06 03:55:46 crc kubenswrapper[4980]: I1206 03:55:46.300835 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/ce3481e1-ecb6-4d13-9258-656e9ac4559d-run\") pod \"glance-default-single-0\" (UID: \"ce3481e1-ecb6-4d13-9258-656e9ac4559d\") " pod="glance-kuttl-tests/glance-default-single-0" Dec 06 03:55:46 crc kubenswrapper[4980]: I1206 03:55:46.301067 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/ce3481e1-ecb6-4d13-9258-656e9ac4559d-etc-iscsi\") pod \"glance-default-single-0\" (UID: \"ce3481e1-ecb6-4d13-9258-656e9ac4559d\") " pod="glance-kuttl-tests/glance-default-single-0" Dec 06 03:55:46 crc kubenswrapper[4980]: I1206 03:55:46.301089 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/ce3481e1-ecb6-4d13-9258-656e9ac4559d-httpd-run\") pod \"glance-default-single-0\" (UID: \"ce3481e1-ecb6-4d13-9258-656e9ac4559d\") " pod="glance-kuttl-tests/glance-default-single-0" Dec 06 03:55:46 crc kubenswrapper[4980]: I1206 03:55:46.301117 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ce3481e1-ecb6-4d13-9258-656e9ac4559d-config-data\") pod \"glance-default-single-0\" (UID: \"ce3481e1-ecb6-4d13-9258-656e9ac4559d\") " pod="glance-kuttl-tests/glance-default-single-0" Dec 06 03:55:46 crc kubenswrapper[4980]: I1206 03:55:46.301165 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/ce3481e1-ecb6-4d13-9258-656e9ac4559d-lib-modules\") pod \"glance-default-single-0\" (UID: \"ce3481e1-ecb6-4d13-9258-656e9ac4559d\") " pod="glance-kuttl-tests/glance-default-single-0" Dec 06 03:55:46 crc kubenswrapper[4980]: I1206 03:55:46.301176 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/ce3481e1-ecb6-4d13-9258-656e9ac4559d-etc-iscsi\") pod \"glance-default-single-0\" (UID: \"ce3481e1-ecb6-4d13-9258-656e9ac4559d\") " pod="glance-kuttl-tests/glance-default-single-0" Dec 06 03:55:46 crc kubenswrapper[4980]: I1206 03:55:46.301191 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/ce3481e1-ecb6-4d13-9258-656e9ac4559d-dev\") pod \"glance-default-single-0\" (UID: \"ce3481e1-ecb6-4d13-9258-656e9ac4559d\") " pod="glance-kuttl-tests/glance-default-single-0" Dec 06 03:55:46 crc kubenswrapper[4980]: I1206 03:55:46.301419 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/ce3481e1-ecb6-4d13-9258-656e9ac4559d-dev\") pod \"glance-default-single-0\" (UID: \"ce3481e1-ecb6-4d13-9258-656e9ac4559d\") " pod="glance-kuttl-tests/glance-default-single-0" Dec 06 03:55:46 crc kubenswrapper[4980]: I1206 03:55:46.301443 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/ce3481e1-ecb6-4d13-9258-656e9ac4559d-httpd-run\") pod \"glance-default-single-0\" (UID: \"ce3481e1-ecb6-4d13-9258-656e9ac4559d\") " pod="glance-kuttl-tests/glance-default-single-0" Dec 06 03:55:46 crc kubenswrapper[4980]: I1206 03:55:46.301465 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/ce3481e1-ecb6-4d13-9258-656e9ac4559d-lib-modules\") pod \"glance-default-single-0\" (UID: \"ce3481e1-ecb6-4d13-9258-656e9ac4559d\") " pod="glance-kuttl-tests/glance-default-single-0" Dec 06 03:55:46 crc kubenswrapper[4980]: I1206 03:55:46.301478 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ce3481e1-ecb6-4d13-9258-656e9ac4559d-logs\") pod \"glance-default-single-0\" (UID: \"ce3481e1-ecb6-4d13-9258-656e9ac4559d\") " pod="glance-kuttl-tests/glance-default-single-0" Dec 06 03:55:46 crc kubenswrapper[4980]: I1206 03:55:46.301537 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/ce3481e1-ecb6-4d13-9258-656e9ac4559d-run\") pod \"glance-default-single-0\" (UID: \"ce3481e1-ecb6-4d13-9258-656e9ac4559d\") " pod="glance-kuttl-tests/glance-default-single-0" Dec 06 03:55:46 crc kubenswrapper[4980]: I1206 03:55:46.301568 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/ce3481e1-ecb6-4d13-9258-656e9ac4559d-sys\") pod \"glance-default-single-0\" (UID: \"ce3481e1-ecb6-4d13-9258-656e9ac4559d\") " pod="glance-kuttl-tests/glance-default-single-0" Dec 06 03:55:46 crc kubenswrapper[4980]: I1206 03:55:46.301545 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/ce3481e1-ecb6-4d13-9258-656e9ac4559d-sys\") pod \"glance-default-single-0\" (UID: \"ce3481e1-ecb6-4d13-9258-656e9ac4559d\") " pod="glance-kuttl-tests/glance-default-single-0" Dec 06 03:55:46 crc kubenswrapper[4980]: I1206 03:55:46.301621 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-single-0\" (UID: \"ce3481e1-ecb6-4d13-9258-656e9ac4559d\") " pod="glance-kuttl-tests/glance-default-single-0" Dec 06 03:55:46 crc kubenswrapper[4980]: I1206 03:55:46.301852 4980 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-single-0\" (UID: \"ce3481e1-ecb6-4d13-9258-656e9ac4559d\") device mount path \"/mnt/openstack/pv05\"" pod="glance-kuttl-tests/glance-default-single-0" Dec 06 03:55:46 crc kubenswrapper[4980]: I1206 03:55:46.306820 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ce3481e1-ecb6-4d13-9258-656e9ac4559d-scripts\") pod \"glance-default-single-0\" (UID: \"ce3481e1-ecb6-4d13-9258-656e9ac4559d\") " pod="glance-kuttl-tests/glance-default-single-0" Dec 06 03:55:46 crc kubenswrapper[4980]: I1206 03:55:46.306944 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ce3481e1-ecb6-4d13-9258-656e9ac4559d-config-data\") pod \"glance-default-single-0\" (UID: \"ce3481e1-ecb6-4d13-9258-656e9ac4559d\") " pod="glance-kuttl-tests/glance-default-single-0" Dec 06 03:55:46 crc kubenswrapper[4980]: I1206 03:55:46.320923 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-single-0\" (UID: \"ce3481e1-ecb6-4d13-9258-656e9ac4559d\") " pod="glance-kuttl-tests/glance-default-single-0" Dec 06 03:55:46 crc kubenswrapper[4980]: I1206 03:55:46.331586 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7rpnn\" (UniqueName: \"kubernetes.io/projected/ce3481e1-ecb6-4d13-9258-656e9ac4559d-kube-api-access-7rpnn\") pod \"glance-default-single-0\" (UID: \"ce3481e1-ecb6-4d13-9258-656e9ac4559d\") " pod="glance-kuttl-tests/glance-default-single-0" Dec 06 03:55:46 crc kubenswrapper[4980]: I1206 03:55:46.346259 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-single-0\" (UID: \"ce3481e1-ecb6-4d13-9258-656e9ac4559d\") " pod="glance-kuttl-tests/glance-default-single-0" Dec 06 03:55:46 crc kubenswrapper[4980]: I1206 03:55:46.392959 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-0" Dec 06 03:55:46 crc kubenswrapper[4980]: I1206 03:55:46.724245 4980 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Dec 06 03:55:47 crc kubenswrapper[4980]: I1206 03:55:47.193287 4980 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="de2e2787-7471-41a3-9bd7-e591410f3f42" path="/var/lib/kubelet/pods/de2e2787-7471-41a3-9bd7-e591410f3f42/volumes" Dec 06 03:55:47 crc kubenswrapper[4980]: I1206 03:55:47.684283 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"ce3481e1-ecb6-4d13-9258-656e9ac4559d","Type":"ContainerStarted","Data":"9810a0efa5bf67def7aafca7fffdd26361ed9ecc47f4b498266a3d9b14086431"} Dec 06 03:55:47 crc kubenswrapper[4980]: I1206 03:55:47.684348 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"ce3481e1-ecb6-4d13-9258-656e9ac4559d","Type":"ContainerStarted","Data":"a94542a921b42102e0ae8a10fa7d7352b68bb5523587c23de62886fdf77c5780"} Dec 06 03:55:47 crc kubenswrapper[4980]: I1206 03:55:47.684367 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"ce3481e1-ecb6-4d13-9258-656e9ac4559d","Type":"ContainerStarted","Data":"bd8b6fbccec4a54f0c223b953a590c69d80909687b77c33fa6771cf3b8bb9865"} Dec 06 03:55:47 crc kubenswrapper[4980]: I1206 03:55:47.716999 4980 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/glance-default-single-0" podStartSLOduration=1.7169537369999999 podStartE2EDuration="1.716953737s" podCreationTimestamp="2025-12-06 03:55:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 03:55:47.712688125 +0000 UTC m=+1306.957439396" watchObservedRunningTime="2025-12-06 03:55:47.716953737 +0000 UTC m=+1306.961705028" Dec 06 03:55:55 crc kubenswrapper[4980]: I1206 03:55:55.137784 4980 patch_prober.go:28] interesting pod/machine-config-daemon-r5zfb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 03:55:55 crc kubenswrapper[4980]: I1206 03:55:55.138429 4980 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-r5zfb" podUID="320f44d1-a671-4a91-b328-a8b0fdd8f23a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 03:55:56 crc kubenswrapper[4980]: I1206 03:55:56.393533 4980 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-single-0" Dec 06 03:55:56 crc kubenswrapper[4980]: I1206 03:55:56.393582 4980 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-single-0" Dec 06 03:55:56 crc kubenswrapper[4980]: I1206 03:55:56.448583 4980 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-single-0" Dec 06 03:55:56 crc kubenswrapper[4980]: I1206 03:55:56.460507 4980 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-single-0" Dec 06 03:55:56 crc kubenswrapper[4980]: I1206 03:55:56.750998 4980 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-single-0" Dec 06 03:55:56 crc kubenswrapper[4980]: I1206 03:55:56.751059 4980 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-single-0" Dec 06 03:55:58 crc kubenswrapper[4980]: I1206 03:55:58.736156 4980 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-single-0" Dec 06 03:55:58 crc kubenswrapper[4980]: I1206 03:55:58.746213 4980 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-single-0" Dec 06 03:56:17 crc kubenswrapper[4980]: I1206 03:56:17.492967 4980 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-db-sync-bdqlq"] Dec 06 03:56:17 crc kubenswrapper[4980]: I1206 03:56:17.502055 4980 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-db-sync-bdqlq"] Dec 06 03:56:17 crc kubenswrapper[4980]: I1206 03:56:17.562405 4980 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glancef836-account-delete-2w2rk"] Dec 06 03:56:17 crc kubenswrapper[4980]: I1206 03:56:17.563965 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glancef836-account-delete-2w2rk" Dec 06 03:56:17 crc kubenswrapper[4980]: I1206 03:56:17.604559 4980 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glancef836-account-delete-2w2rk"] Dec 06 03:56:17 crc kubenswrapper[4980]: I1206 03:56:17.611337 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6g7nc\" (UniqueName: \"kubernetes.io/projected/aa8eabbe-3442-45c6-9b53-34f10ab6c39f-kube-api-access-6g7nc\") pod \"glancef836-account-delete-2w2rk\" (UID: \"aa8eabbe-3442-45c6-9b53-34f10ab6c39f\") " pod="glance-kuttl-tests/glancef836-account-delete-2w2rk" Dec 06 03:56:17 crc kubenswrapper[4980]: I1206 03:56:17.626091 4980 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Dec 06 03:56:17 crc kubenswrapper[4980]: I1206 03:56:17.626356 4980 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-single-0" podUID="ce3481e1-ecb6-4d13-9258-656e9ac4559d" containerName="glance-log" containerID="cri-o://a94542a921b42102e0ae8a10fa7d7352b68bb5523587c23de62886fdf77c5780" gracePeriod=30 Dec 06 03:56:17 crc kubenswrapper[4980]: I1206 03:56:17.626503 4980 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-single-0" podUID="ce3481e1-ecb6-4d13-9258-656e9ac4559d" containerName="glance-httpd" containerID="cri-o://9810a0efa5bf67def7aafca7fffdd26361ed9ecc47f4b498266a3d9b14086431" gracePeriod=30 Dec 06 03:56:17 crc kubenswrapper[4980]: I1206 03:56:17.642053 4980 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-single-1"] Dec 06 03:56:17 crc kubenswrapper[4980]: I1206 03:56:17.642327 4980 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-single-1" podUID="5bfed592-a73b-4218-bd4a-eab04765a748" containerName="glance-log" containerID="cri-o://b89a93f06cb4564f65b336ba76c926fcf3f7cf40cc9bf58d7af24dd958e2a086" gracePeriod=30 Dec 06 03:56:17 crc kubenswrapper[4980]: I1206 03:56:17.642810 4980 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-single-1" podUID="5bfed592-a73b-4218-bd4a-eab04765a748" containerName="glance-httpd" containerID="cri-o://ccfd68de53a61d766d5202a884a95ac78a14bdcda5e296b2a9ece932bde45d0a" gracePeriod=30 Dec 06 03:56:17 crc kubenswrapper[4980]: I1206 03:56:17.713347 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6g7nc\" (UniqueName: \"kubernetes.io/projected/aa8eabbe-3442-45c6-9b53-34f10ab6c39f-kube-api-access-6g7nc\") pod \"glancef836-account-delete-2w2rk\" (UID: \"aa8eabbe-3442-45c6-9b53-34f10ab6c39f\") " pod="glance-kuttl-tests/glancef836-account-delete-2w2rk" Dec 06 03:56:17 crc kubenswrapper[4980]: I1206 03:56:17.734320 4980 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/openstackclient"] Dec 06 03:56:17 crc kubenswrapper[4980]: I1206 03:56:17.734616 4980 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/openstackclient" podUID="7c853f1b-b0d8-4f38-9ad3-0dcb3b2e0707" containerName="openstackclient" containerID="cri-o://5a6c6907ac4004409028d10f1c5c0615205912ae630e496046598525b59ccad4" gracePeriod=30 Dec 06 03:56:17 crc kubenswrapper[4980]: I1206 03:56:17.742256 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6g7nc\" (UniqueName: \"kubernetes.io/projected/aa8eabbe-3442-45c6-9b53-34f10ab6c39f-kube-api-access-6g7nc\") pod \"glancef836-account-delete-2w2rk\" (UID: \"aa8eabbe-3442-45c6-9b53-34f10ab6c39f\") " pod="glance-kuttl-tests/glancef836-account-delete-2w2rk" Dec 06 03:56:17 crc kubenswrapper[4980]: I1206 03:56:17.900907 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glancef836-account-delete-2w2rk" Dec 06 03:56:17 crc kubenswrapper[4980]: I1206 03:56:17.959950 4980 generic.go:334] "Generic (PLEG): container finished" podID="5bfed592-a73b-4218-bd4a-eab04765a748" containerID="b89a93f06cb4564f65b336ba76c926fcf3f7cf40cc9bf58d7af24dd958e2a086" exitCode=143 Dec 06 03:56:17 crc kubenswrapper[4980]: I1206 03:56:17.960020 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-1" event={"ID":"5bfed592-a73b-4218-bd4a-eab04765a748","Type":"ContainerDied","Data":"b89a93f06cb4564f65b336ba76c926fcf3f7cf40cc9bf58d7af24dd958e2a086"} Dec 06 03:56:17 crc kubenswrapper[4980]: I1206 03:56:17.973861 4980 generic.go:334] "Generic (PLEG): container finished" podID="ce3481e1-ecb6-4d13-9258-656e9ac4559d" containerID="a94542a921b42102e0ae8a10fa7d7352b68bb5523587c23de62886fdf77c5780" exitCode=143 Dec 06 03:56:17 crc kubenswrapper[4980]: I1206 03:56:17.973976 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"ce3481e1-ecb6-4d13-9258-656e9ac4559d","Type":"ContainerDied","Data":"a94542a921b42102e0ae8a10fa7d7352b68bb5523587c23de62886fdf77c5780"} Dec 06 03:56:17 crc kubenswrapper[4980]: I1206 03:56:17.976360 4980 generic.go:334] "Generic (PLEG): container finished" podID="7c853f1b-b0d8-4f38-9ad3-0dcb3b2e0707" containerID="5a6c6907ac4004409028d10f1c5c0615205912ae630e496046598525b59ccad4" exitCode=143 Dec 06 03:56:17 crc kubenswrapper[4980]: I1206 03:56:17.976397 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/openstackclient" event={"ID":"7c853f1b-b0d8-4f38-9ad3-0dcb3b2e0707","Type":"ContainerDied","Data":"5a6c6907ac4004409028d10f1c5c0615205912ae630e496046598525b59ccad4"} Dec 06 03:56:18 crc kubenswrapper[4980]: I1206 03:56:18.150943 4980 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/openstackclient" Dec 06 03:56:18 crc kubenswrapper[4980]: I1206 03:56:18.193959 4980 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glancef836-account-delete-2w2rk"] Dec 06 03:56:18 crc kubenswrapper[4980]: I1206 03:56:18.321870 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/7c853f1b-b0d8-4f38-9ad3-0dcb3b2e0707-openstack-config\") pod \"7c853f1b-b0d8-4f38-9ad3-0dcb3b2e0707\" (UID: \"7c853f1b-b0d8-4f38-9ad3-0dcb3b2e0707\") " Dec 06 03:56:18 crc kubenswrapper[4980]: I1206 03:56:18.321931 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-scripts\" (UniqueName: \"kubernetes.io/configmap/7c853f1b-b0d8-4f38-9ad3-0dcb3b2e0707-openstack-scripts\") pod \"7c853f1b-b0d8-4f38-9ad3-0dcb3b2e0707\" (UID: \"7c853f1b-b0d8-4f38-9ad3-0dcb3b2e0707\") " Dec 06 03:56:18 crc kubenswrapper[4980]: I1206 03:56:18.322814 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7c853f1b-b0d8-4f38-9ad3-0dcb3b2e0707-openstack-scripts" (OuterVolumeSpecName: "openstack-scripts") pod "7c853f1b-b0d8-4f38-9ad3-0dcb3b2e0707" (UID: "7c853f1b-b0d8-4f38-9ad3-0dcb3b2e0707"). InnerVolumeSpecName "openstack-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 03:56:18 crc kubenswrapper[4980]: I1206 03:56:18.322875 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5rrbt\" (UniqueName: \"kubernetes.io/projected/7c853f1b-b0d8-4f38-9ad3-0dcb3b2e0707-kube-api-access-5rrbt\") pod \"7c853f1b-b0d8-4f38-9ad3-0dcb3b2e0707\" (UID: \"7c853f1b-b0d8-4f38-9ad3-0dcb3b2e0707\") " Dec 06 03:56:18 crc kubenswrapper[4980]: I1206 03:56:18.323282 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/7c853f1b-b0d8-4f38-9ad3-0dcb3b2e0707-openstack-config-secret\") pod \"7c853f1b-b0d8-4f38-9ad3-0dcb3b2e0707\" (UID: \"7c853f1b-b0d8-4f38-9ad3-0dcb3b2e0707\") " Dec 06 03:56:18 crc kubenswrapper[4980]: I1206 03:56:18.323569 4980 reconciler_common.go:293] "Volume detached for volume \"openstack-scripts\" (UniqueName: \"kubernetes.io/configmap/7c853f1b-b0d8-4f38-9ad3-0dcb3b2e0707-openstack-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 03:56:18 crc kubenswrapper[4980]: I1206 03:56:18.326267 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7c853f1b-b0d8-4f38-9ad3-0dcb3b2e0707-kube-api-access-5rrbt" (OuterVolumeSpecName: "kube-api-access-5rrbt") pod "7c853f1b-b0d8-4f38-9ad3-0dcb3b2e0707" (UID: "7c853f1b-b0d8-4f38-9ad3-0dcb3b2e0707"). InnerVolumeSpecName "kube-api-access-5rrbt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 03:56:18 crc kubenswrapper[4980]: I1206 03:56:18.341595 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7c853f1b-b0d8-4f38-9ad3-0dcb3b2e0707-openstack-config" (OuterVolumeSpecName: "openstack-config") pod "7c853f1b-b0d8-4f38-9ad3-0dcb3b2e0707" (UID: "7c853f1b-b0d8-4f38-9ad3-0dcb3b2e0707"). InnerVolumeSpecName "openstack-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 03:56:18 crc kubenswrapper[4980]: I1206 03:56:18.345583 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7c853f1b-b0d8-4f38-9ad3-0dcb3b2e0707-openstack-config-secret" (OuterVolumeSpecName: "openstack-config-secret") pod "7c853f1b-b0d8-4f38-9ad3-0dcb3b2e0707" (UID: "7c853f1b-b0d8-4f38-9ad3-0dcb3b2e0707"). InnerVolumeSpecName "openstack-config-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 03:56:18 crc kubenswrapper[4980]: I1206 03:56:18.424484 4980 reconciler_common.go:293] "Volume detached for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/7c853f1b-b0d8-4f38-9ad3-0dcb3b2e0707-openstack-config-secret\") on node \"crc\" DevicePath \"\"" Dec 06 03:56:18 crc kubenswrapper[4980]: I1206 03:56:18.424543 4980 reconciler_common.go:293] "Volume detached for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/7c853f1b-b0d8-4f38-9ad3-0dcb3b2e0707-openstack-config\") on node \"crc\" DevicePath \"\"" Dec 06 03:56:18 crc kubenswrapper[4980]: I1206 03:56:18.424560 4980 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5rrbt\" (UniqueName: \"kubernetes.io/projected/7c853f1b-b0d8-4f38-9ad3-0dcb3b2e0707-kube-api-access-5rrbt\") on node \"crc\" DevicePath \"\"" Dec 06 03:56:18 crc kubenswrapper[4980]: I1206 03:56:18.984962 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/openstackclient" event={"ID":"7c853f1b-b0d8-4f38-9ad3-0dcb3b2e0707","Type":"ContainerDied","Data":"ef6423d15175d50287cb4ccb582811f976e235b548f1c363508a945cd94d83c7"} Dec 06 03:56:18 crc kubenswrapper[4980]: I1206 03:56:18.985440 4980 scope.go:117] "RemoveContainer" containerID="5a6c6907ac4004409028d10f1c5c0615205912ae630e496046598525b59ccad4" Dec 06 03:56:18 crc kubenswrapper[4980]: I1206 03:56:18.985336 4980 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/openstackclient" Dec 06 03:56:18 crc kubenswrapper[4980]: I1206 03:56:18.986971 4980 generic.go:334] "Generic (PLEG): container finished" podID="aa8eabbe-3442-45c6-9b53-34f10ab6c39f" containerID="753730fa2418548cbcb343d5fd27fc6d2f49bbf8d8505e21c4025c5fcd2dc732" exitCode=0 Dec 06 03:56:18 crc kubenswrapper[4980]: I1206 03:56:18.987004 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glancef836-account-delete-2w2rk" event={"ID":"aa8eabbe-3442-45c6-9b53-34f10ab6c39f","Type":"ContainerDied","Data":"753730fa2418548cbcb343d5fd27fc6d2f49bbf8d8505e21c4025c5fcd2dc732"} Dec 06 03:56:18 crc kubenswrapper[4980]: I1206 03:56:18.987041 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glancef836-account-delete-2w2rk" event={"ID":"aa8eabbe-3442-45c6-9b53-34f10ab6c39f","Type":"ContainerStarted","Data":"224f141f256de3d84718723e4fd94b3a0ea4c32ec2bf6f703ebb3ba8e3cbaaa7"} Dec 06 03:56:19 crc kubenswrapper[4980]: I1206 03:56:19.044357 4980 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/openstackclient"] Dec 06 03:56:19 crc kubenswrapper[4980]: I1206 03:56:19.050263 4980 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/openstackclient"] Dec 06 03:56:19 crc kubenswrapper[4980]: I1206 03:56:19.196045 4980 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7c853f1b-b0d8-4f38-9ad3-0dcb3b2e0707" path="/var/lib/kubelet/pods/7c853f1b-b0d8-4f38-9ad3-0dcb3b2e0707/volumes" Dec 06 03:56:19 crc kubenswrapper[4980]: I1206 03:56:19.196852 4980 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d8db6f70-28bc-4b58-91e9-feb5bd06a3e4" path="/var/lib/kubelet/pods/d8db6f70-28bc-4b58-91e9-feb5bd06a3e4/volumes" Dec 06 03:56:20 crc kubenswrapper[4980]: I1206 03:56:20.264493 4980 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glancef836-account-delete-2w2rk" Dec 06 03:56:20 crc kubenswrapper[4980]: I1206 03:56:20.454054 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6g7nc\" (UniqueName: \"kubernetes.io/projected/aa8eabbe-3442-45c6-9b53-34f10ab6c39f-kube-api-access-6g7nc\") pod \"aa8eabbe-3442-45c6-9b53-34f10ab6c39f\" (UID: \"aa8eabbe-3442-45c6-9b53-34f10ab6c39f\") " Dec 06 03:56:20 crc kubenswrapper[4980]: I1206 03:56:20.460083 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/aa8eabbe-3442-45c6-9b53-34f10ab6c39f-kube-api-access-6g7nc" (OuterVolumeSpecName: "kube-api-access-6g7nc") pod "aa8eabbe-3442-45c6-9b53-34f10ab6c39f" (UID: "aa8eabbe-3442-45c6-9b53-34f10ab6c39f"). InnerVolumeSpecName "kube-api-access-6g7nc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 03:56:20 crc kubenswrapper[4980]: I1206 03:56:20.556054 4980 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6g7nc\" (UniqueName: \"kubernetes.io/projected/aa8eabbe-3442-45c6-9b53-34f10ab6c39f-kube-api-access-6g7nc\") on node \"crc\" DevicePath \"\"" Dec 06 03:56:21 crc kubenswrapper[4980]: I1206 03:56:21.017551 4980 generic.go:334] "Generic (PLEG): container finished" podID="ce3481e1-ecb6-4d13-9258-656e9ac4559d" containerID="9810a0efa5bf67def7aafca7fffdd26361ed9ecc47f4b498266a3d9b14086431" exitCode=0 Dec 06 03:56:21 crc kubenswrapper[4980]: I1206 03:56:21.017628 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"ce3481e1-ecb6-4d13-9258-656e9ac4559d","Type":"ContainerDied","Data":"9810a0efa5bf67def7aafca7fffdd26361ed9ecc47f4b498266a3d9b14086431"} Dec 06 03:56:21 crc kubenswrapper[4980]: I1206 03:56:21.020582 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glancef836-account-delete-2w2rk" event={"ID":"aa8eabbe-3442-45c6-9b53-34f10ab6c39f","Type":"ContainerDied","Data":"224f141f256de3d84718723e4fd94b3a0ea4c32ec2bf6f703ebb3ba8e3cbaaa7"} Dec 06 03:56:21 crc kubenswrapper[4980]: I1206 03:56:21.020610 4980 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="224f141f256de3d84718723e4fd94b3a0ea4c32ec2bf6f703ebb3ba8e3cbaaa7" Dec 06 03:56:21 crc kubenswrapper[4980]: I1206 03:56:21.020634 4980 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glancef836-account-delete-2w2rk" Dec 06 03:56:21 crc kubenswrapper[4980]: I1206 03:56:21.045238 4980 generic.go:334] "Generic (PLEG): container finished" podID="5bfed592-a73b-4218-bd4a-eab04765a748" containerID="ccfd68de53a61d766d5202a884a95ac78a14bdcda5e296b2a9ece932bde45d0a" exitCode=0 Dec 06 03:56:21 crc kubenswrapper[4980]: I1206 03:56:21.045278 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-1" event={"ID":"5bfed592-a73b-4218-bd4a-eab04765a748","Type":"ContainerDied","Data":"ccfd68de53a61d766d5202a884a95ac78a14bdcda5e296b2a9ece932bde45d0a"} Dec 06 03:56:21 crc kubenswrapper[4980]: I1206 03:56:21.318794 4980 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-0" Dec 06 03:56:21 crc kubenswrapper[4980]: I1206 03:56:21.323060 4980 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-1" Dec 06 03:56:21 crc kubenswrapper[4980]: I1206 03:56:21.391148 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/ce3481e1-ecb6-4d13-9258-656e9ac4559d-sys\") pod \"ce3481e1-ecb6-4d13-9258-656e9ac4559d\" (UID: \"ce3481e1-ecb6-4d13-9258-656e9ac4559d\") " Dec 06 03:56:21 crc kubenswrapper[4980]: I1206 03:56:21.391194 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/5bfed592-a73b-4218-bd4a-eab04765a748-sys\") pod \"5bfed592-a73b-4218-bd4a-eab04765a748\" (UID: \"5bfed592-a73b-4218-bd4a-eab04765a748\") " Dec 06 03:56:21 crc kubenswrapper[4980]: I1206 03:56:21.391237 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5bfed592-a73b-4218-bd4a-eab04765a748-scripts\") pod \"5bfed592-a73b-4218-bd4a-eab04765a748\" (UID: \"5bfed592-a73b-4218-bd4a-eab04765a748\") " Dec 06 03:56:21 crc kubenswrapper[4980]: I1206 03:56:21.391253 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/ce3481e1-ecb6-4d13-9258-656e9ac4559d-dev\") pod \"ce3481e1-ecb6-4d13-9258-656e9ac4559d\" (UID: \"ce3481e1-ecb6-4d13-9258-656e9ac4559d\") " Dec 06 03:56:21 crc kubenswrapper[4980]: I1206 03:56:21.391277 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance-cache\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"ce3481e1-ecb6-4d13-9258-656e9ac4559d\" (UID: \"ce3481e1-ecb6-4d13-9258-656e9ac4559d\") " Dec 06 03:56:21 crc kubenswrapper[4980]: I1206 03:56:21.391313 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/5bfed592-a73b-4218-bd4a-eab04765a748-lib-modules\") pod \"5bfed592-a73b-4218-bd4a-eab04765a748\" (UID: \"5bfed592-a73b-4218-bd4a-eab04765a748\") " Dec 06 03:56:21 crc kubenswrapper[4980]: I1206 03:56:21.391335 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/5bfed592-a73b-4218-bd4a-eab04765a748-dev\") pod \"5bfed592-a73b-4218-bd4a-eab04765a748\" (UID: \"5bfed592-a73b-4218-bd4a-eab04765a748\") " Dec 06 03:56:21 crc kubenswrapper[4980]: I1206 03:56:21.391331 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/5bfed592-a73b-4218-bd4a-eab04765a748-sys" (OuterVolumeSpecName: "sys") pod "5bfed592-a73b-4218-bd4a-eab04765a748" (UID: "5bfed592-a73b-4218-bd4a-eab04765a748"). InnerVolumeSpecName "sys". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 03:56:21 crc kubenswrapper[4980]: I1206 03:56:21.391351 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5bfed592-a73b-4218-bd4a-eab04765a748-config-data\") pod \"5bfed592-a73b-4218-bd4a-eab04765a748\" (UID: \"5bfed592-a73b-4218-bd4a-eab04765a748\") " Dec 06 03:56:21 crc kubenswrapper[4980]: I1206 03:56:21.391421 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/5bfed592-a73b-4218-bd4a-eab04765a748-run\") pod \"5bfed592-a73b-4218-bd4a-eab04765a748\" (UID: \"5bfed592-a73b-4218-bd4a-eab04765a748\") " Dec 06 03:56:21 crc kubenswrapper[4980]: I1206 03:56:21.391452 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/ce3481e1-ecb6-4d13-9258-656e9ac4559d-httpd-run\") pod \"ce3481e1-ecb6-4d13-9258-656e9ac4559d\" (UID: \"ce3481e1-ecb6-4d13-9258-656e9ac4559d\") " Dec 06 03:56:21 crc kubenswrapper[4980]: I1206 03:56:21.391475 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/ce3481e1-ecb6-4d13-9258-656e9ac4559d-etc-iscsi\") pod \"ce3481e1-ecb6-4d13-9258-656e9ac4559d\" (UID: \"ce3481e1-ecb6-4d13-9258-656e9ac4559d\") " Dec 06 03:56:21 crc kubenswrapper[4980]: I1206 03:56:21.391494 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5bfed592-a73b-4218-bd4a-eab04765a748-logs\") pod \"5bfed592-a73b-4218-bd4a-eab04765a748\" (UID: \"5bfed592-a73b-4218-bd4a-eab04765a748\") " Dec 06 03:56:21 crc kubenswrapper[4980]: I1206 03:56:21.391538 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/5bfed592-a73b-4218-bd4a-eab04765a748-var-locks-brick\") pod \"5bfed592-a73b-4218-bd4a-eab04765a748\" (UID: \"5bfed592-a73b-4218-bd4a-eab04765a748\") " Dec 06 03:56:21 crc kubenswrapper[4980]: I1206 03:56:21.391556 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ce3481e1-ecb6-4d13-9258-656e9ac4559d-config-data\") pod \"ce3481e1-ecb6-4d13-9258-656e9ac4559d\" (UID: \"ce3481e1-ecb6-4d13-9258-656e9ac4559d\") " Dec 06 03:56:21 crc kubenswrapper[4980]: I1206 03:56:21.391573 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7rpnn\" (UniqueName: \"kubernetes.io/projected/ce3481e1-ecb6-4d13-9258-656e9ac4559d-kube-api-access-7rpnn\") pod \"ce3481e1-ecb6-4d13-9258-656e9ac4559d\" (UID: \"ce3481e1-ecb6-4d13-9258-656e9ac4559d\") " Dec 06 03:56:21 crc kubenswrapper[4980]: I1206 03:56:21.391602 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ce3481e1-ecb6-4d13-9258-656e9ac4559d-logs\") pod \"ce3481e1-ecb6-4d13-9258-656e9ac4559d\" (UID: \"ce3481e1-ecb6-4d13-9258-656e9ac4559d\") " Dec 06 03:56:21 crc kubenswrapper[4980]: I1206 03:56:21.391625 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/5bfed592-a73b-4218-bd4a-eab04765a748-etc-iscsi\") pod \"5bfed592-a73b-4218-bd4a-eab04765a748\" (UID: \"5bfed592-a73b-4218-bd4a-eab04765a748\") " Dec 06 03:56:21 crc kubenswrapper[4980]: I1206 03:56:21.391645 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/ce3481e1-ecb6-4d13-9258-656e9ac4559d-etc-nvme\") pod \"ce3481e1-ecb6-4d13-9258-656e9ac4559d\" (UID: \"ce3481e1-ecb6-4d13-9258-656e9ac4559d\") " Dec 06 03:56:21 crc kubenswrapper[4980]: I1206 03:56:21.391677 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"5bfed592-a73b-4218-bd4a-eab04765a748\" (UID: \"5bfed592-a73b-4218-bd4a-eab04765a748\") " Dec 06 03:56:21 crc kubenswrapper[4980]: I1206 03:56:21.391692 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/5bfed592-a73b-4218-bd4a-eab04765a748-etc-nvme\") pod \"5bfed592-a73b-4218-bd4a-eab04765a748\" (UID: \"5bfed592-a73b-4218-bd4a-eab04765a748\") " Dec 06 03:56:21 crc kubenswrapper[4980]: I1206 03:56:21.391712 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tvc5b\" (UniqueName: \"kubernetes.io/projected/5bfed592-a73b-4218-bd4a-eab04765a748-kube-api-access-tvc5b\") pod \"5bfed592-a73b-4218-bd4a-eab04765a748\" (UID: \"5bfed592-a73b-4218-bd4a-eab04765a748\") " Dec 06 03:56:21 crc kubenswrapper[4980]: I1206 03:56:21.391726 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance-cache\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"5bfed592-a73b-4218-bd4a-eab04765a748\" (UID: \"5bfed592-a73b-4218-bd4a-eab04765a748\") " Dec 06 03:56:21 crc kubenswrapper[4980]: I1206 03:56:21.391752 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ce3481e1-ecb6-4d13-9258-656e9ac4559d-scripts\") pod \"ce3481e1-ecb6-4d13-9258-656e9ac4559d\" (UID: \"ce3481e1-ecb6-4d13-9258-656e9ac4559d\") " Dec 06 03:56:21 crc kubenswrapper[4980]: I1206 03:56:21.391780 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/5bfed592-a73b-4218-bd4a-eab04765a748-lib-modules" (OuterVolumeSpecName: "lib-modules") pod "5bfed592-a73b-4218-bd4a-eab04765a748" (UID: "5bfed592-a73b-4218-bd4a-eab04765a748"). InnerVolumeSpecName "lib-modules". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 03:56:21 crc kubenswrapper[4980]: I1206 03:56:21.391836 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ce3481e1-ecb6-4d13-9258-656e9ac4559d-dev" (OuterVolumeSpecName: "dev") pod "ce3481e1-ecb6-4d13-9258-656e9ac4559d" (UID: "ce3481e1-ecb6-4d13-9258-656e9ac4559d"). InnerVolumeSpecName "dev". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 03:56:21 crc kubenswrapper[4980]: I1206 03:56:21.391851 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/5bfed592-a73b-4218-bd4a-eab04765a748-var-locks-brick" (OuterVolumeSpecName: "var-locks-brick") pod "5bfed592-a73b-4218-bd4a-eab04765a748" (UID: "5bfed592-a73b-4218-bd4a-eab04765a748"). InnerVolumeSpecName "var-locks-brick". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 03:56:21 crc kubenswrapper[4980]: I1206 03:56:21.391890 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/5bfed592-a73b-4218-bd4a-eab04765a748-run" (OuterVolumeSpecName: "run") pod "5bfed592-a73b-4218-bd4a-eab04765a748" (UID: "5bfed592-a73b-4218-bd4a-eab04765a748"). InnerVolumeSpecName "run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 03:56:21 crc kubenswrapper[4980]: I1206 03:56:21.391906 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ce3481e1-ecb6-4d13-9258-656e9ac4559d-etc-iscsi" (OuterVolumeSpecName: "etc-iscsi") pod "ce3481e1-ecb6-4d13-9258-656e9ac4559d" (UID: "ce3481e1-ecb6-4d13-9258-656e9ac4559d"). InnerVolumeSpecName "etc-iscsi". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 03:56:21 crc kubenswrapper[4980]: I1206 03:56:21.392108 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/5bfed592-a73b-4218-bd4a-eab04765a748-etc-iscsi" (OuterVolumeSpecName: "etc-iscsi") pod "5bfed592-a73b-4218-bd4a-eab04765a748" (UID: "5bfed592-a73b-4218-bd4a-eab04765a748"). InnerVolumeSpecName "etc-iscsi". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 03:56:21 crc kubenswrapper[4980]: I1206 03:56:21.392185 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/5bfed592-a73b-4218-bd4a-eab04765a748-dev" (OuterVolumeSpecName: "dev") pod "5bfed592-a73b-4218-bd4a-eab04765a748" (UID: "5bfed592-a73b-4218-bd4a-eab04765a748"). InnerVolumeSpecName "dev". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 03:56:21 crc kubenswrapper[4980]: I1206 03:56:21.392240 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ce3481e1-ecb6-4d13-9258-656e9ac4559d-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "ce3481e1-ecb6-4d13-9258-656e9ac4559d" (UID: "ce3481e1-ecb6-4d13-9258-656e9ac4559d"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 03:56:21 crc kubenswrapper[4980]: I1206 03:56:21.392267 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5bfed592-a73b-4218-bd4a-eab04765a748-logs" (OuterVolumeSpecName: "logs") pod "5bfed592-a73b-4218-bd4a-eab04765a748" (UID: "5bfed592-a73b-4218-bd4a-eab04765a748"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 03:56:21 crc kubenswrapper[4980]: I1206 03:56:21.392311 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ce3481e1-ecb6-4d13-9258-656e9ac4559d-etc-nvme" (OuterVolumeSpecName: "etc-nvme") pod "ce3481e1-ecb6-4d13-9258-656e9ac4559d" (UID: "ce3481e1-ecb6-4d13-9258-656e9ac4559d"). InnerVolumeSpecName "etc-nvme". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 03:56:21 crc kubenswrapper[4980]: I1206 03:56:21.392315 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/5bfed592-a73b-4218-bd4a-eab04765a748-etc-nvme" (OuterVolumeSpecName: "etc-nvme") pod "5bfed592-a73b-4218-bd4a-eab04765a748" (UID: "5bfed592-a73b-4218-bd4a-eab04765a748"). InnerVolumeSpecName "etc-nvme". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 03:56:21 crc kubenswrapper[4980]: I1206 03:56:21.392418 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ce3481e1-ecb6-4d13-9258-656e9ac4559d-logs" (OuterVolumeSpecName: "logs") pod "ce3481e1-ecb6-4d13-9258-656e9ac4559d" (UID: "ce3481e1-ecb6-4d13-9258-656e9ac4559d"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 03:56:21 crc kubenswrapper[4980]: I1206 03:56:21.392478 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/ce3481e1-ecb6-4d13-9258-656e9ac4559d-lib-modules\") pod \"ce3481e1-ecb6-4d13-9258-656e9ac4559d\" (UID: \"ce3481e1-ecb6-4d13-9258-656e9ac4559d\") " Dec 06 03:56:21 crc kubenswrapper[4980]: I1206 03:56:21.392544 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ce3481e1-ecb6-4d13-9258-656e9ac4559d-lib-modules" (OuterVolumeSpecName: "lib-modules") pod "ce3481e1-ecb6-4d13-9258-656e9ac4559d" (UID: "ce3481e1-ecb6-4d13-9258-656e9ac4559d"). InnerVolumeSpecName "lib-modules". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 03:56:21 crc kubenswrapper[4980]: I1206 03:56:21.392609 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/5bfed592-a73b-4218-bd4a-eab04765a748-httpd-run\") pod \"5bfed592-a73b-4218-bd4a-eab04765a748\" (UID: \"5bfed592-a73b-4218-bd4a-eab04765a748\") " Dec 06 03:56:21 crc kubenswrapper[4980]: I1206 03:56:21.392628 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"ce3481e1-ecb6-4d13-9258-656e9ac4559d\" (UID: \"ce3481e1-ecb6-4d13-9258-656e9ac4559d\") " Dec 06 03:56:21 crc kubenswrapper[4980]: I1206 03:56:21.392642 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/ce3481e1-ecb6-4d13-9258-656e9ac4559d-run\") pod \"ce3481e1-ecb6-4d13-9258-656e9ac4559d\" (UID: \"ce3481e1-ecb6-4d13-9258-656e9ac4559d\") " Dec 06 03:56:21 crc kubenswrapper[4980]: I1206 03:56:21.392668 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/ce3481e1-ecb6-4d13-9258-656e9ac4559d-var-locks-brick\") pod \"ce3481e1-ecb6-4d13-9258-656e9ac4559d\" (UID: \"ce3481e1-ecb6-4d13-9258-656e9ac4559d\") " Dec 06 03:56:21 crc kubenswrapper[4980]: I1206 03:56:21.393143 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ce3481e1-ecb6-4d13-9258-656e9ac4559d-sys" (OuterVolumeSpecName: "sys") pod "ce3481e1-ecb6-4d13-9258-656e9ac4559d" (UID: "ce3481e1-ecb6-4d13-9258-656e9ac4559d"). InnerVolumeSpecName "sys". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 03:56:21 crc kubenswrapper[4980]: I1206 03:56:21.393225 4980 reconciler_common.go:293] "Volume detached for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/ce3481e1-ecb6-4d13-9258-656e9ac4559d-etc-nvme\") on node \"crc\" DevicePath \"\"" Dec 06 03:56:21 crc kubenswrapper[4980]: I1206 03:56:21.393240 4980 reconciler_common.go:293] "Volume detached for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/5bfed592-a73b-4218-bd4a-eab04765a748-etc-nvme\") on node \"crc\" DevicePath \"\"" Dec 06 03:56:21 crc kubenswrapper[4980]: I1206 03:56:21.393276 4980 reconciler_common.go:293] "Volume detached for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/ce3481e1-ecb6-4d13-9258-656e9ac4559d-lib-modules\") on node \"crc\" DevicePath \"\"" Dec 06 03:56:21 crc kubenswrapper[4980]: I1206 03:56:21.393288 4980 reconciler_common.go:293] "Volume detached for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/5bfed592-a73b-4218-bd4a-eab04765a748-sys\") on node \"crc\" DevicePath \"\"" Dec 06 03:56:21 crc kubenswrapper[4980]: I1206 03:56:21.393296 4980 reconciler_common.go:293] "Volume detached for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/ce3481e1-ecb6-4d13-9258-656e9ac4559d-dev\") on node \"crc\" DevicePath \"\"" Dec 06 03:56:21 crc kubenswrapper[4980]: I1206 03:56:21.393304 4980 reconciler_common.go:293] "Volume detached for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/5bfed592-a73b-4218-bd4a-eab04765a748-lib-modules\") on node \"crc\" DevicePath \"\"" Dec 06 03:56:21 crc kubenswrapper[4980]: I1206 03:56:21.393311 4980 reconciler_common.go:293] "Volume detached for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/5bfed592-a73b-4218-bd4a-eab04765a748-dev\") on node \"crc\" DevicePath \"\"" Dec 06 03:56:21 crc kubenswrapper[4980]: I1206 03:56:21.393319 4980 reconciler_common.go:293] "Volume detached for volume \"run\" (UniqueName: \"kubernetes.io/host-path/5bfed592-a73b-4218-bd4a-eab04765a748-run\") on node \"crc\" DevicePath \"\"" Dec 06 03:56:21 crc kubenswrapper[4980]: I1206 03:56:21.393326 4980 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/ce3481e1-ecb6-4d13-9258-656e9ac4559d-httpd-run\") on node \"crc\" DevicePath \"\"" Dec 06 03:56:21 crc kubenswrapper[4980]: I1206 03:56:21.393333 4980 reconciler_common.go:293] "Volume detached for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/ce3481e1-ecb6-4d13-9258-656e9ac4559d-etc-iscsi\") on node \"crc\" DevicePath \"\"" Dec 06 03:56:21 crc kubenswrapper[4980]: I1206 03:56:21.393342 4980 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5bfed592-a73b-4218-bd4a-eab04765a748-logs\") on node \"crc\" DevicePath \"\"" Dec 06 03:56:21 crc kubenswrapper[4980]: I1206 03:56:21.393350 4980 reconciler_common.go:293] "Volume detached for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/5bfed592-a73b-4218-bd4a-eab04765a748-var-locks-brick\") on node \"crc\" DevicePath \"\"" Dec 06 03:56:21 crc kubenswrapper[4980]: I1206 03:56:21.393359 4980 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ce3481e1-ecb6-4d13-9258-656e9ac4559d-logs\") on node \"crc\" DevicePath \"\"" Dec 06 03:56:21 crc kubenswrapper[4980]: I1206 03:56:21.393366 4980 reconciler_common.go:293] "Volume detached for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/5bfed592-a73b-4218-bd4a-eab04765a748-etc-iscsi\") on node \"crc\" DevicePath \"\"" Dec 06 03:56:21 crc kubenswrapper[4980]: I1206 03:56:21.393392 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ce3481e1-ecb6-4d13-9258-656e9ac4559d-var-locks-brick" (OuterVolumeSpecName: "var-locks-brick") pod "ce3481e1-ecb6-4d13-9258-656e9ac4559d" (UID: "ce3481e1-ecb6-4d13-9258-656e9ac4559d"). InnerVolumeSpecName "var-locks-brick". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 03:56:21 crc kubenswrapper[4980]: I1206 03:56:21.393413 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ce3481e1-ecb6-4d13-9258-656e9ac4559d-run" (OuterVolumeSpecName: "run") pod "ce3481e1-ecb6-4d13-9258-656e9ac4559d" (UID: "ce3481e1-ecb6-4d13-9258-656e9ac4559d"). InnerVolumeSpecName "run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 03:56:21 crc kubenswrapper[4980]: I1206 03:56:21.393751 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5bfed592-a73b-4218-bd4a-eab04765a748-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "5bfed592-a73b-4218-bd4a-eab04765a748" (UID: "5bfed592-a73b-4218-bd4a-eab04765a748"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 03:56:21 crc kubenswrapper[4980]: I1206 03:56:21.396707 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage07-crc" (OuterVolumeSpecName: "glance-cache") pod "5bfed592-a73b-4218-bd4a-eab04765a748" (UID: "5bfed592-a73b-4218-bd4a-eab04765a748"). InnerVolumeSpecName "local-storage07-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 06 03:56:21 crc kubenswrapper[4980]: I1206 03:56:21.396722 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ce3481e1-ecb6-4d13-9258-656e9ac4559d-scripts" (OuterVolumeSpecName: "scripts") pod "ce3481e1-ecb6-4d13-9258-656e9ac4559d" (UID: "ce3481e1-ecb6-4d13-9258-656e9ac4559d"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 03:56:21 crc kubenswrapper[4980]: I1206 03:56:21.396715 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage05-crc" (OuterVolumeSpecName: "glance-cache") pod "ce3481e1-ecb6-4d13-9258-656e9ac4559d" (UID: "ce3481e1-ecb6-4d13-9258-656e9ac4559d"). InnerVolumeSpecName "local-storage05-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 06 03:56:21 crc kubenswrapper[4980]: I1206 03:56:21.396732 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5bfed592-a73b-4218-bd4a-eab04765a748-scripts" (OuterVolumeSpecName: "scripts") pod "5bfed592-a73b-4218-bd4a-eab04765a748" (UID: "5bfed592-a73b-4218-bd4a-eab04765a748"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 03:56:21 crc kubenswrapper[4980]: I1206 03:56:21.396793 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ce3481e1-ecb6-4d13-9258-656e9ac4559d-kube-api-access-7rpnn" (OuterVolumeSpecName: "kube-api-access-7rpnn") pod "ce3481e1-ecb6-4d13-9258-656e9ac4559d" (UID: "ce3481e1-ecb6-4d13-9258-656e9ac4559d"). InnerVolumeSpecName "kube-api-access-7rpnn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 03:56:21 crc kubenswrapper[4980]: I1206 03:56:21.398181 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage08-crc" (OuterVolumeSpecName: "glance") pod "5bfed592-a73b-4218-bd4a-eab04765a748" (UID: "5bfed592-a73b-4218-bd4a-eab04765a748"). InnerVolumeSpecName "local-storage08-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 06 03:56:21 crc kubenswrapper[4980]: I1206 03:56:21.399478 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage09-crc" (OuterVolumeSpecName: "glance") pod "ce3481e1-ecb6-4d13-9258-656e9ac4559d" (UID: "ce3481e1-ecb6-4d13-9258-656e9ac4559d"). InnerVolumeSpecName "local-storage09-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 06 03:56:21 crc kubenswrapper[4980]: I1206 03:56:21.411022 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5bfed592-a73b-4218-bd4a-eab04765a748-kube-api-access-tvc5b" (OuterVolumeSpecName: "kube-api-access-tvc5b") pod "5bfed592-a73b-4218-bd4a-eab04765a748" (UID: "5bfed592-a73b-4218-bd4a-eab04765a748"). InnerVolumeSpecName "kube-api-access-tvc5b". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 03:56:21 crc kubenswrapper[4980]: I1206 03:56:21.432648 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ce3481e1-ecb6-4d13-9258-656e9ac4559d-config-data" (OuterVolumeSpecName: "config-data") pod "ce3481e1-ecb6-4d13-9258-656e9ac4559d" (UID: "ce3481e1-ecb6-4d13-9258-656e9ac4559d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 03:56:21 crc kubenswrapper[4980]: I1206 03:56:21.435847 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5bfed592-a73b-4218-bd4a-eab04765a748-config-data" (OuterVolumeSpecName: "config-data") pod "5bfed592-a73b-4218-bd4a-eab04765a748" (UID: "5bfed592-a73b-4218-bd4a-eab04765a748"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 03:56:21 crc kubenswrapper[4980]: I1206 03:56:21.495031 4980 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5bfed592-a73b-4218-bd4a-eab04765a748-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 03:56:21 crc kubenswrapper[4980]: I1206 03:56:21.495099 4980 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") on node \"crc\" " Dec 06 03:56:21 crc kubenswrapper[4980]: I1206 03:56:21.495113 4980 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5bfed592-a73b-4218-bd4a-eab04765a748-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 03:56:21 crc kubenswrapper[4980]: I1206 03:56:21.495122 4980 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ce3481e1-ecb6-4d13-9258-656e9ac4559d-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 03:56:21 crc kubenswrapper[4980]: I1206 03:56:21.495131 4980 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7rpnn\" (UniqueName: \"kubernetes.io/projected/ce3481e1-ecb6-4d13-9258-656e9ac4559d-kube-api-access-7rpnn\") on node \"crc\" DevicePath \"\"" Dec 06 03:56:21 crc kubenswrapper[4980]: I1206 03:56:21.495149 4980 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") on node \"crc\" " Dec 06 03:56:21 crc kubenswrapper[4980]: I1206 03:56:21.495159 4980 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tvc5b\" (UniqueName: \"kubernetes.io/projected/5bfed592-a73b-4218-bd4a-eab04765a748-kube-api-access-tvc5b\") on node \"crc\" DevicePath \"\"" Dec 06 03:56:21 crc kubenswrapper[4980]: I1206 03:56:21.495174 4980 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") on node \"crc\" " Dec 06 03:56:21 crc kubenswrapper[4980]: I1206 03:56:21.495182 4980 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ce3481e1-ecb6-4d13-9258-656e9ac4559d-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 03:56:21 crc kubenswrapper[4980]: I1206 03:56:21.495190 4980 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/5bfed592-a73b-4218-bd4a-eab04765a748-httpd-run\") on node \"crc\" DevicePath \"\"" Dec 06 03:56:21 crc kubenswrapper[4980]: I1206 03:56:21.495202 4980 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") on node \"crc\" " Dec 06 03:56:21 crc kubenswrapper[4980]: I1206 03:56:21.495211 4980 reconciler_common.go:293] "Volume detached for volume \"run\" (UniqueName: \"kubernetes.io/host-path/ce3481e1-ecb6-4d13-9258-656e9ac4559d-run\") on node \"crc\" DevicePath \"\"" Dec 06 03:56:21 crc kubenswrapper[4980]: I1206 03:56:21.495218 4980 reconciler_common.go:293] "Volume detached for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/ce3481e1-ecb6-4d13-9258-656e9ac4559d-var-locks-brick\") on node \"crc\" DevicePath \"\"" Dec 06 03:56:21 crc kubenswrapper[4980]: I1206 03:56:21.495227 4980 reconciler_common.go:293] "Volume detached for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/ce3481e1-ecb6-4d13-9258-656e9ac4559d-sys\") on node \"crc\" DevicePath \"\"" Dec 06 03:56:21 crc kubenswrapper[4980]: I1206 03:56:21.507125 4980 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage09-crc" (UniqueName: "kubernetes.io/local-volume/local-storage09-crc") on node "crc" Dec 06 03:56:21 crc kubenswrapper[4980]: I1206 03:56:21.508378 4980 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage05-crc" (UniqueName: "kubernetes.io/local-volume/local-storage05-crc") on node "crc" Dec 06 03:56:21 crc kubenswrapper[4980]: I1206 03:56:21.510102 4980 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage08-crc" (UniqueName: "kubernetes.io/local-volume/local-storage08-crc") on node "crc" Dec 06 03:56:21 crc kubenswrapper[4980]: I1206 03:56:21.517256 4980 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage07-crc" (UniqueName: "kubernetes.io/local-volume/local-storage07-crc") on node "crc" Dec 06 03:56:21 crc kubenswrapper[4980]: I1206 03:56:21.596047 4980 reconciler_common.go:293] "Volume detached for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") on node \"crc\" DevicePath \"\"" Dec 06 03:56:21 crc kubenswrapper[4980]: I1206 03:56:21.596089 4980 reconciler_common.go:293] "Volume detached for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") on node \"crc\" DevicePath \"\"" Dec 06 03:56:21 crc kubenswrapper[4980]: I1206 03:56:21.596102 4980 reconciler_common.go:293] "Volume detached for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") on node \"crc\" DevicePath \"\"" Dec 06 03:56:21 crc kubenswrapper[4980]: I1206 03:56:21.596115 4980 reconciler_common.go:293] "Volume detached for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") on node \"crc\" DevicePath \"\"" Dec 06 03:56:22 crc kubenswrapper[4980]: I1206 03:56:22.056460 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-1" event={"ID":"5bfed592-a73b-4218-bd4a-eab04765a748","Type":"ContainerDied","Data":"96c7c655066bf28e24a9711be703ae81dfa33e701fdf7f5874d9586601dc3bab"} Dec 06 03:56:22 crc kubenswrapper[4980]: I1206 03:56:22.056535 4980 scope.go:117] "RemoveContainer" containerID="ccfd68de53a61d766d5202a884a95ac78a14bdcda5e296b2a9ece932bde45d0a" Dec 06 03:56:22 crc kubenswrapper[4980]: I1206 03:56:22.056846 4980 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-1" Dec 06 03:56:22 crc kubenswrapper[4980]: I1206 03:56:22.058488 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"ce3481e1-ecb6-4d13-9258-656e9ac4559d","Type":"ContainerDied","Data":"bd8b6fbccec4a54f0c223b953a590c69d80909687b77c33fa6771cf3b8bb9865"} Dec 06 03:56:22 crc kubenswrapper[4980]: I1206 03:56:22.058555 4980 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-0" Dec 06 03:56:22 crc kubenswrapper[4980]: I1206 03:56:22.092794 4980 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-single-1"] Dec 06 03:56:22 crc kubenswrapper[4980]: I1206 03:56:22.095076 4980 scope.go:117] "RemoveContainer" containerID="b89a93f06cb4564f65b336ba76c926fcf3f7cf40cc9bf58d7af24dd958e2a086" Dec 06 03:56:22 crc kubenswrapper[4980]: I1206 03:56:22.100238 4980 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-default-single-1"] Dec 06 03:56:22 crc kubenswrapper[4980]: I1206 03:56:22.107185 4980 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Dec 06 03:56:22 crc kubenswrapper[4980]: I1206 03:56:22.119224 4980 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Dec 06 03:56:22 crc kubenswrapper[4980]: I1206 03:56:22.131455 4980 scope.go:117] "RemoveContainer" containerID="9810a0efa5bf67def7aafca7fffdd26361ed9ecc47f4b498266a3d9b14086431" Dec 06 03:56:22 crc kubenswrapper[4980]: I1206 03:56:22.151207 4980 scope.go:117] "RemoveContainer" containerID="a94542a921b42102e0ae8a10fa7d7352b68bb5523587c23de62886fdf77c5780" Dec 06 03:56:22 crc kubenswrapper[4980]: I1206 03:56:22.787024 4980 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-db-create-whc99"] Dec 06 03:56:22 crc kubenswrapper[4980]: I1206 03:56:22.793324 4980 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-db-create-whc99"] Dec 06 03:56:22 crc kubenswrapper[4980]: I1206 03:56:22.799702 4980 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glancef836-account-delete-2w2rk"] Dec 06 03:56:22 crc kubenswrapper[4980]: I1206 03:56:22.804929 4980 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-f836-account-create-9h4q6"] Dec 06 03:56:22 crc kubenswrapper[4980]: I1206 03:56:22.810803 4980 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glancef836-account-delete-2w2rk"] Dec 06 03:56:22 crc kubenswrapper[4980]: I1206 03:56:22.817640 4980 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-f836-account-create-9h4q6"] Dec 06 03:56:23 crc kubenswrapper[4980]: I1206 03:56:23.192618 4980 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="50b592be-1d9d-45c7-9cc9-4c7082d2af78" path="/var/lib/kubelet/pods/50b592be-1d9d-45c7-9cc9-4c7082d2af78/volumes" Dec 06 03:56:23 crc kubenswrapper[4980]: I1206 03:56:23.193472 4980 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5bfed592-a73b-4218-bd4a-eab04765a748" path="/var/lib/kubelet/pods/5bfed592-a73b-4218-bd4a-eab04765a748/volumes" Dec 06 03:56:23 crc kubenswrapper[4980]: I1206 03:56:23.194188 4980 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7575d3d1-f9b7-4840-bca9-30d6eaa5d0c3" path="/var/lib/kubelet/pods/7575d3d1-f9b7-4840-bca9-30d6eaa5d0c3/volumes" Dec 06 03:56:23 crc kubenswrapper[4980]: I1206 03:56:23.196017 4980 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="aa8eabbe-3442-45c6-9b53-34f10ab6c39f" path="/var/lib/kubelet/pods/aa8eabbe-3442-45c6-9b53-34f10ab6c39f/volumes" Dec 06 03:56:23 crc kubenswrapper[4980]: I1206 03:56:23.196977 4980 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ce3481e1-ecb6-4d13-9258-656e9ac4559d" path="/var/lib/kubelet/pods/ce3481e1-ecb6-4d13-9258-656e9ac4559d/volumes" Dec 06 03:56:23 crc kubenswrapper[4980]: I1206 03:56:23.370356 4980 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-db-create-nxfng"] Dec 06 03:56:23 crc kubenswrapper[4980]: E1206 03:56:23.370716 4980 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ce3481e1-ecb6-4d13-9258-656e9ac4559d" containerName="glance-log" Dec 06 03:56:23 crc kubenswrapper[4980]: I1206 03:56:23.370799 4980 state_mem.go:107] "Deleted CPUSet assignment" podUID="ce3481e1-ecb6-4d13-9258-656e9ac4559d" containerName="glance-log" Dec 06 03:56:23 crc kubenswrapper[4980]: E1206 03:56:23.370821 4980 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aa8eabbe-3442-45c6-9b53-34f10ab6c39f" containerName="mariadb-account-delete" Dec 06 03:56:23 crc kubenswrapper[4980]: I1206 03:56:23.370829 4980 state_mem.go:107] "Deleted CPUSet assignment" podUID="aa8eabbe-3442-45c6-9b53-34f10ab6c39f" containerName="mariadb-account-delete" Dec 06 03:56:23 crc kubenswrapper[4980]: E1206 03:56:23.370844 4980 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5bfed592-a73b-4218-bd4a-eab04765a748" containerName="glance-log" Dec 06 03:56:23 crc kubenswrapper[4980]: I1206 03:56:23.370853 4980 state_mem.go:107] "Deleted CPUSet assignment" podUID="5bfed592-a73b-4218-bd4a-eab04765a748" containerName="glance-log" Dec 06 03:56:23 crc kubenswrapper[4980]: E1206 03:56:23.370871 4980 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ce3481e1-ecb6-4d13-9258-656e9ac4559d" containerName="glance-httpd" Dec 06 03:56:23 crc kubenswrapper[4980]: I1206 03:56:23.370879 4980 state_mem.go:107] "Deleted CPUSet assignment" podUID="ce3481e1-ecb6-4d13-9258-656e9ac4559d" containerName="glance-httpd" Dec 06 03:56:23 crc kubenswrapper[4980]: E1206 03:56:23.370901 4980 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7c853f1b-b0d8-4f38-9ad3-0dcb3b2e0707" containerName="openstackclient" Dec 06 03:56:23 crc kubenswrapper[4980]: I1206 03:56:23.370914 4980 state_mem.go:107] "Deleted CPUSet assignment" podUID="7c853f1b-b0d8-4f38-9ad3-0dcb3b2e0707" containerName="openstackclient" Dec 06 03:56:23 crc kubenswrapper[4980]: E1206 03:56:23.370934 4980 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5bfed592-a73b-4218-bd4a-eab04765a748" containerName="glance-httpd" Dec 06 03:56:23 crc kubenswrapper[4980]: I1206 03:56:23.370944 4980 state_mem.go:107] "Deleted CPUSet assignment" podUID="5bfed592-a73b-4218-bd4a-eab04765a748" containerName="glance-httpd" Dec 06 03:56:23 crc kubenswrapper[4980]: I1206 03:56:23.371201 4980 memory_manager.go:354] "RemoveStaleState removing state" podUID="aa8eabbe-3442-45c6-9b53-34f10ab6c39f" containerName="mariadb-account-delete" Dec 06 03:56:23 crc kubenswrapper[4980]: I1206 03:56:23.371221 4980 memory_manager.go:354] "RemoveStaleState removing state" podUID="7c853f1b-b0d8-4f38-9ad3-0dcb3b2e0707" containerName="openstackclient" Dec 06 03:56:23 crc kubenswrapper[4980]: I1206 03:56:23.371235 4980 memory_manager.go:354] "RemoveStaleState removing state" podUID="ce3481e1-ecb6-4d13-9258-656e9ac4559d" containerName="glance-log" Dec 06 03:56:23 crc kubenswrapper[4980]: I1206 03:56:23.371253 4980 memory_manager.go:354] "RemoveStaleState removing state" podUID="5bfed592-a73b-4218-bd4a-eab04765a748" containerName="glance-log" Dec 06 03:56:23 crc kubenswrapper[4980]: I1206 03:56:23.371269 4980 memory_manager.go:354] "RemoveStaleState removing state" podUID="5bfed592-a73b-4218-bd4a-eab04765a748" containerName="glance-httpd" Dec 06 03:56:23 crc kubenswrapper[4980]: I1206 03:56:23.371283 4980 memory_manager.go:354] "RemoveStaleState removing state" podUID="ce3481e1-ecb6-4d13-9258-656e9ac4559d" containerName="glance-httpd" Dec 06 03:56:23 crc kubenswrapper[4980]: I1206 03:56:23.371991 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-create-nxfng" Dec 06 03:56:23 crc kubenswrapper[4980]: I1206 03:56:23.393824 4980 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-db-create-nxfng"] Dec 06 03:56:23 crc kubenswrapper[4980]: I1206 03:56:23.561734 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ztcf5\" (UniqueName: \"kubernetes.io/projected/d3e30c7b-9b62-40e0-9fca-773b737214a9-kube-api-access-ztcf5\") pod \"glance-db-create-nxfng\" (UID: \"d3e30c7b-9b62-40e0-9fca-773b737214a9\") " pod="glance-kuttl-tests/glance-db-create-nxfng" Dec 06 03:56:23 crc kubenswrapper[4980]: I1206 03:56:23.663545 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ztcf5\" (UniqueName: \"kubernetes.io/projected/d3e30c7b-9b62-40e0-9fca-773b737214a9-kube-api-access-ztcf5\") pod \"glance-db-create-nxfng\" (UID: \"d3e30c7b-9b62-40e0-9fca-773b737214a9\") " pod="glance-kuttl-tests/glance-db-create-nxfng" Dec 06 03:56:23 crc kubenswrapper[4980]: I1206 03:56:23.685328 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ztcf5\" (UniqueName: \"kubernetes.io/projected/d3e30c7b-9b62-40e0-9fca-773b737214a9-kube-api-access-ztcf5\") pod \"glance-db-create-nxfng\" (UID: \"d3e30c7b-9b62-40e0-9fca-773b737214a9\") " pod="glance-kuttl-tests/glance-db-create-nxfng" Dec 06 03:56:23 crc kubenswrapper[4980]: I1206 03:56:23.709122 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-create-nxfng" Dec 06 03:56:24 crc kubenswrapper[4980]: I1206 03:56:24.176067 4980 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-db-create-nxfng"] Dec 06 03:56:24 crc kubenswrapper[4980]: W1206 03:56:24.185032 4980 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd3e30c7b_9b62_40e0_9fca_773b737214a9.slice/crio-e78751156e5827308ccafc7b45e826307e52e617789fa6dd2a526bf1fc9ea04e WatchSource:0}: Error finding container e78751156e5827308ccafc7b45e826307e52e617789fa6dd2a526bf1fc9ea04e: Status 404 returned error can't find the container with id e78751156e5827308ccafc7b45e826307e52e617789fa6dd2a526bf1fc9ea04e Dec 06 03:56:25 crc kubenswrapper[4980]: I1206 03:56:25.087698 4980 generic.go:334] "Generic (PLEG): container finished" podID="d3e30c7b-9b62-40e0-9fca-773b737214a9" containerID="6c4578e685304543174fec33d1b046ae66eb3d061795db4c986db1a947718b35" exitCode=0 Dec 06 03:56:25 crc kubenswrapper[4980]: I1206 03:56:25.087748 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-create-nxfng" event={"ID":"d3e30c7b-9b62-40e0-9fca-773b737214a9","Type":"ContainerDied","Data":"6c4578e685304543174fec33d1b046ae66eb3d061795db4c986db1a947718b35"} Dec 06 03:56:25 crc kubenswrapper[4980]: I1206 03:56:25.088115 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-create-nxfng" event={"ID":"d3e30c7b-9b62-40e0-9fca-773b737214a9","Type":"ContainerStarted","Data":"e78751156e5827308ccafc7b45e826307e52e617789fa6dd2a526bf1fc9ea04e"} Dec 06 03:56:25 crc kubenswrapper[4980]: I1206 03:56:25.137662 4980 patch_prober.go:28] interesting pod/machine-config-daemon-r5zfb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 03:56:25 crc kubenswrapper[4980]: I1206 03:56:25.137733 4980 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-r5zfb" podUID="320f44d1-a671-4a91-b328-a8b0fdd8f23a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 03:56:25 crc kubenswrapper[4980]: I1206 03:56:25.137782 4980 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-r5zfb" Dec 06 03:56:25 crc kubenswrapper[4980]: I1206 03:56:25.138582 4980 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"ed3956e825d000e94b97025c20ede41bab44553dd92465ad73c08a9ec44729eb"} pod="openshift-machine-config-operator/machine-config-daemon-r5zfb" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 06 03:56:25 crc kubenswrapper[4980]: I1206 03:56:25.138671 4980 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-r5zfb" podUID="320f44d1-a671-4a91-b328-a8b0fdd8f23a" containerName="machine-config-daemon" containerID="cri-o://ed3956e825d000e94b97025c20ede41bab44553dd92465ad73c08a9ec44729eb" gracePeriod=600 Dec 06 03:56:26 crc kubenswrapper[4980]: I1206 03:56:26.117741 4980 generic.go:334] "Generic (PLEG): container finished" podID="320f44d1-a671-4a91-b328-a8b0fdd8f23a" containerID="ed3956e825d000e94b97025c20ede41bab44553dd92465ad73c08a9ec44729eb" exitCode=0 Dec 06 03:56:26 crc kubenswrapper[4980]: I1206 03:56:26.117844 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-r5zfb" event={"ID":"320f44d1-a671-4a91-b328-a8b0fdd8f23a","Type":"ContainerDied","Data":"ed3956e825d000e94b97025c20ede41bab44553dd92465ad73c08a9ec44729eb"} Dec 06 03:56:26 crc kubenswrapper[4980]: I1206 03:56:26.118505 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-r5zfb" event={"ID":"320f44d1-a671-4a91-b328-a8b0fdd8f23a","Type":"ContainerStarted","Data":"e8878e13c398cceec46bb52c4c9ea0e06dbf4317e6ad39b69ca763a0d8b15497"} Dec 06 03:56:26 crc kubenswrapper[4980]: I1206 03:56:26.118614 4980 scope.go:117] "RemoveContainer" containerID="a9db17110188a0ff97c6851d4bf74c82d9f9edeef9b7cfcb0fd15528f7d2e059" Dec 06 03:56:26 crc kubenswrapper[4980]: I1206 03:56:26.457706 4980 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-create-nxfng" Dec 06 03:56:26 crc kubenswrapper[4980]: I1206 03:56:26.612766 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ztcf5\" (UniqueName: \"kubernetes.io/projected/d3e30c7b-9b62-40e0-9fca-773b737214a9-kube-api-access-ztcf5\") pod \"d3e30c7b-9b62-40e0-9fca-773b737214a9\" (UID: \"d3e30c7b-9b62-40e0-9fca-773b737214a9\") " Dec 06 03:56:26 crc kubenswrapper[4980]: I1206 03:56:26.617943 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d3e30c7b-9b62-40e0-9fca-773b737214a9-kube-api-access-ztcf5" (OuterVolumeSpecName: "kube-api-access-ztcf5") pod "d3e30c7b-9b62-40e0-9fca-773b737214a9" (UID: "d3e30c7b-9b62-40e0-9fca-773b737214a9"). InnerVolumeSpecName "kube-api-access-ztcf5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 03:56:26 crc kubenswrapper[4980]: I1206 03:56:26.713898 4980 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ztcf5\" (UniqueName: \"kubernetes.io/projected/d3e30c7b-9b62-40e0-9fca-773b737214a9-kube-api-access-ztcf5\") on node \"crc\" DevicePath \"\"" Dec 06 03:56:27 crc kubenswrapper[4980]: I1206 03:56:27.129216 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-create-nxfng" event={"ID":"d3e30c7b-9b62-40e0-9fca-773b737214a9","Type":"ContainerDied","Data":"e78751156e5827308ccafc7b45e826307e52e617789fa6dd2a526bf1fc9ea04e"} Dec 06 03:56:27 crc kubenswrapper[4980]: I1206 03:56:27.129264 4980 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e78751156e5827308ccafc7b45e826307e52e617789fa6dd2a526bf1fc9ea04e" Dec 06 03:56:27 crc kubenswrapper[4980]: I1206 03:56:27.129268 4980 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-create-nxfng" Dec 06 03:56:33 crc kubenswrapper[4980]: I1206 03:56:33.416645 4980 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-970a-account-create-8b64q"] Dec 06 03:56:33 crc kubenswrapper[4980]: E1206 03:56:33.417425 4980 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d3e30c7b-9b62-40e0-9fca-773b737214a9" containerName="mariadb-database-create" Dec 06 03:56:33 crc kubenswrapper[4980]: I1206 03:56:33.417447 4980 state_mem.go:107] "Deleted CPUSet assignment" podUID="d3e30c7b-9b62-40e0-9fca-773b737214a9" containerName="mariadb-database-create" Dec 06 03:56:33 crc kubenswrapper[4980]: I1206 03:56:33.417730 4980 memory_manager.go:354] "RemoveStaleState removing state" podUID="d3e30c7b-9b62-40e0-9fca-773b737214a9" containerName="mariadb-database-create" Dec 06 03:56:33 crc kubenswrapper[4980]: I1206 03:56:33.418397 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-970a-account-create-8b64q" Dec 06 03:56:33 crc kubenswrapper[4980]: I1206 03:56:33.422261 4980 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-db-secret" Dec 06 03:56:33 crc kubenswrapper[4980]: I1206 03:56:33.431541 4980 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-970a-account-create-8b64q"] Dec 06 03:56:33 crc kubenswrapper[4980]: I1206 03:56:33.568278 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f62p9\" (UniqueName: \"kubernetes.io/projected/bb09c398-429b-4870-a06b-0ebe969f814e-kube-api-access-f62p9\") pod \"glance-970a-account-create-8b64q\" (UID: \"bb09c398-429b-4870-a06b-0ebe969f814e\") " pod="glance-kuttl-tests/glance-970a-account-create-8b64q" Dec 06 03:56:33 crc kubenswrapper[4980]: I1206 03:56:33.669919 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f62p9\" (UniqueName: \"kubernetes.io/projected/bb09c398-429b-4870-a06b-0ebe969f814e-kube-api-access-f62p9\") pod \"glance-970a-account-create-8b64q\" (UID: \"bb09c398-429b-4870-a06b-0ebe969f814e\") " pod="glance-kuttl-tests/glance-970a-account-create-8b64q" Dec 06 03:56:33 crc kubenswrapper[4980]: I1206 03:56:33.699602 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f62p9\" (UniqueName: \"kubernetes.io/projected/bb09c398-429b-4870-a06b-0ebe969f814e-kube-api-access-f62p9\") pod \"glance-970a-account-create-8b64q\" (UID: \"bb09c398-429b-4870-a06b-0ebe969f814e\") " pod="glance-kuttl-tests/glance-970a-account-create-8b64q" Dec 06 03:56:33 crc kubenswrapper[4980]: I1206 03:56:33.746436 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-970a-account-create-8b64q" Dec 06 03:56:34 crc kubenswrapper[4980]: I1206 03:56:34.200222 4980 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-970a-account-create-8b64q"] Dec 06 03:56:35 crc kubenswrapper[4980]: I1206 03:56:35.198526 4980 generic.go:334] "Generic (PLEG): container finished" podID="bb09c398-429b-4870-a06b-0ebe969f814e" containerID="8951fad3dafe21d4c37c7ed72d986cd9c42fbc4bd08df4cce7c8c43f2e7a4625" exitCode=0 Dec 06 03:56:35 crc kubenswrapper[4980]: I1206 03:56:35.198610 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-970a-account-create-8b64q" event={"ID":"bb09c398-429b-4870-a06b-0ebe969f814e","Type":"ContainerDied","Data":"8951fad3dafe21d4c37c7ed72d986cd9c42fbc4bd08df4cce7c8c43f2e7a4625"} Dec 06 03:56:35 crc kubenswrapper[4980]: I1206 03:56:35.198651 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-970a-account-create-8b64q" event={"ID":"bb09c398-429b-4870-a06b-0ebe969f814e","Type":"ContainerStarted","Data":"380eadc67a173b10456f5e4a8297477b23a0dde1ba10b284f84f3188ca8d8704"} Dec 06 03:56:36 crc kubenswrapper[4980]: I1206 03:56:36.482614 4980 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-970a-account-create-8b64q" Dec 06 03:56:36 crc kubenswrapper[4980]: I1206 03:56:36.622681 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f62p9\" (UniqueName: \"kubernetes.io/projected/bb09c398-429b-4870-a06b-0ebe969f814e-kube-api-access-f62p9\") pod \"bb09c398-429b-4870-a06b-0ebe969f814e\" (UID: \"bb09c398-429b-4870-a06b-0ebe969f814e\") " Dec 06 03:56:36 crc kubenswrapper[4980]: I1206 03:56:36.637805 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bb09c398-429b-4870-a06b-0ebe969f814e-kube-api-access-f62p9" (OuterVolumeSpecName: "kube-api-access-f62p9") pod "bb09c398-429b-4870-a06b-0ebe969f814e" (UID: "bb09c398-429b-4870-a06b-0ebe969f814e"). InnerVolumeSpecName "kube-api-access-f62p9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 03:56:36 crc kubenswrapper[4980]: I1206 03:56:36.725193 4980 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f62p9\" (UniqueName: \"kubernetes.io/projected/bb09c398-429b-4870-a06b-0ebe969f814e-kube-api-access-f62p9\") on node \"crc\" DevicePath \"\"" Dec 06 03:56:37 crc kubenswrapper[4980]: I1206 03:56:37.218185 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-970a-account-create-8b64q" event={"ID":"bb09c398-429b-4870-a06b-0ebe969f814e","Type":"ContainerDied","Data":"380eadc67a173b10456f5e4a8297477b23a0dde1ba10b284f84f3188ca8d8704"} Dec 06 03:56:37 crc kubenswrapper[4980]: I1206 03:56:37.218585 4980 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="380eadc67a173b10456f5e4a8297477b23a0dde1ba10b284f84f3188ca8d8704" Dec 06 03:56:37 crc kubenswrapper[4980]: I1206 03:56:37.218499 4980 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-970a-account-create-8b64q" Dec 06 03:56:38 crc kubenswrapper[4980]: I1206 03:56:38.635368 4980 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-db-sync-5lft5"] Dec 06 03:56:38 crc kubenswrapper[4980]: E1206 03:56:38.635730 4980 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bb09c398-429b-4870-a06b-0ebe969f814e" containerName="mariadb-account-create" Dec 06 03:56:38 crc kubenswrapper[4980]: I1206 03:56:38.635747 4980 state_mem.go:107] "Deleted CPUSet assignment" podUID="bb09c398-429b-4870-a06b-0ebe969f814e" containerName="mariadb-account-create" Dec 06 03:56:38 crc kubenswrapper[4980]: I1206 03:56:38.635892 4980 memory_manager.go:354] "RemoveStaleState removing state" podUID="bb09c398-429b-4870-a06b-0ebe969f814e" containerName="mariadb-account-create" Dec 06 03:56:38 crc kubenswrapper[4980]: I1206 03:56:38.636481 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-sync-5lft5" Dec 06 03:56:38 crc kubenswrapper[4980]: I1206 03:56:38.638852 4980 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"combined-ca-bundle" Dec 06 03:56:38 crc kubenswrapper[4980]: I1206 03:56:38.639057 4980 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-glance-dockercfg-kjjjr" Dec 06 03:56:38 crc kubenswrapper[4980]: I1206 03:56:38.639365 4980 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-config-data" Dec 06 03:56:38 crc kubenswrapper[4980]: I1206 03:56:38.652023 4980 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-db-sync-5lft5"] Dec 06 03:56:38 crc kubenswrapper[4980]: I1206 03:56:38.770994 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/fca3de68-d822-49ad-bf87-9b0fab269bbb-db-sync-config-data\") pod \"glance-db-sync-5lft5\" (UID: \"fca3de68-d822-49ad-bf87-9b0fab269bbb\") " pod="glance-kuttl-tests/glance-db-sync-5lft5" Dec 06 03:56:38 crc kubenswrapper[4980]: I1206 03:56:38.771206 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fca3de68-d822-49ad-bf87-9b0fab269bbb-config-data\") pod \"glance-db-sync-5lft5\" (UID: \"fca3de68-d822-49ad-bf87-9b0fab269bbb\") " pod="glance-kuttl-tests/glance-db-sync-5lft5" Dec 06 03:56:38 crc kubenswrapper[4980]: I1206 03:56:38.771253 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fca3de68-d822-49ad-bf87-9b0fab269bbb-combined-ca-bundle\") pod \"glance-db-sync-5lft5\" (UID: \"fca3de68-d822-49ad-bf87-9b0fab269bbb\") " pod="glance-kuttl-tests/glance-db-sync-5lft5" Dec 06 03:56:38 crc kubenswrapper[4980]: I1206 03:56:38.771294 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hvfng\" (UniqueName: \"kubernetes.io/projected/fca3de68-d822-49ad-bf87-9b0fab269bbb-kube-api-access-hvfng\") pod \"glance-db-sync-5lft5\" (UID: \"fca3de68-d822-49ad-bf87-9b0fab269bbb\") " pod="glance-kuttl-tests/glance-db-sync-5lft5" Dec 06 03:56:38 crc kubenswrapper[4980]: I1206 03:56:38.872259 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fca3de68-d822-49ad-bf87-9b0fab269bbb-config-data\") pod \"glance-db-sync-5lft5\" (UID: \"fca3de68-d822-49ad-bf87-9b0fab269bbb\") " pod="glance-kuttl-tests/glance-db-sync-5lft5" Dec 06 03:56:38 crc kubenswrapper[4980]: I1206 03:56:38.872313 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fca3de68-d822-49ad-bf87-9b0fab269bbb-combined-ca-bundle\") pod \"glance-db-sync-5lft5\" (UID: \"fca3de68-d822-49ad-bf87-9b0fab269bbb\") " pod="glance-kuttl-tests/glance-db-sync-5lft5" Dec 06 03:56:38 crc kubenswrapper[4980]: I1206 03:56:38.872339 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hvfng\" (UniqueName: \"kubernetes.io/projected/fca3de68-d822-49ad-bf87-9b0fab269bbb-kube-api-access-hvfng\") pod \"glance-db-sync-5lft5\" (UID: \"fca3de68-d822-49ad-bf87-9b0fab269bbb\") " pod="glance-kuttl-tests/glance-db-sync-5lft5" Dec 06 03:56:38 crc kubenswrapper[4980]: I1206 03:56:38.872404 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/fca3de68-d822-49ad-bf87-9b0fab269bbb-db-sync-config-data\") pod \"glance-db-sync-5lft5\" (UID: \"fca3de68-d822-49ad-bf87-9b0fab269bbb\") " pod="glance-kuttl-tests/glance-db-sync-5lft5" Dec 06 03:56:38 crc kubenswrapper[4980]: I1206 03:56:38.881215 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fca3de68-d822-49ad-bf87-9b0fab269bbb-combined-ca-bundle\") pod \"glance-db-sync-5lft5\" (UID: \"fca3de68-d822-49ad-bf87-9b0fab269bbb\") " pod="glance-kuttl-tests/glance-db-sync-5lft5" Dec 06 03:56:38 crc kubenswrapper[4980]: I1206 03:56:38.881280 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/fca3de68-d822-49ad-bf87-9b0fab269bbb-db-sync-config-data\") pod \"glance-db-sync-5lft5\" (UID: \"fca3de68-d822-49ad-bf87-9b0fab269bbb\") " pod="glance-kuttl-tests/glance-db-sync-5lft5" Dec 06 03:56:38 crc kubenswrapper[4980]: I1206 03:56:38.883450 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fca3de68-d822-49ad-bf87-9b0fab269bbb-config-data\") pod \"glance-db-sync-5lft5\" (UID: \"fca3de68-d822-49ad-bf87-9b0fab269bbb\") " pod="glance-kuttl-tests/glance-db-sync-5lft5" Dec 06 03:56:38 crc kubenswrapper[4980]: I1206 03:56:38.900072 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hvfng\" (UniqueName: \"kubernetes.io/projected/fca3de68-d822-49ad-bf87-9b0fab269bbb-kube-api-access-hvfng\") pod \"glance-db-sync-5lft5\" (UID: \"fca3de68-d822-49ad-bf87-9b0fab269bbb\") " pod="glance-kuttl-tests/glance-db-sync-5lft5" Dec 06 03:56:38 crc kubenswrapper[4980]: I1206 03:56:38.952160 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-sync-5lft5" Dec 06 03:56:39 crc kubenswrapper[4980]: I1206 03:56:39.183306 4980 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-db-sync-5lft5"] Dec 06 03:56:39 crc kubenswrapper[4980]: I1206 03:56:39.237168 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-sync-5lft5" event={"ID":"fca3de68-d822-49ad-bf87-9b0fab269bbb","Type":"ContainerStarted","Data":"4a6f918015ca3b061c127fdf679308be47f35733e60b4651a4bbea0d5c492b2d"} Dec 06 03:56:40 crc kubenswrapper[4980]: I1206 03:56:40.244426 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-sync-5lft5" event={"ID":"fca3de68-d822-49ad-bf87-9b0fab269bbb","Type":"ContainerStarted","Data":"fb73adf57f7fa906fd21886e3d9c5da1211cee117c1e8b663adb5a31071287d8"} Dec 06 03:56:40 crc kubenswrapper[4980]: I1206 03:56:40.263471 4980 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/glance-db-sync-5lft5" podStartSLOduration=2.263435884 podStartE2EDuration="2.263435884s" podCreationTimestamp="2025-12-06 03:56:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 03:56:40.259290385 +0000 UTC m=+1359.504041656" watchObservedRunningTime="2025-12-06 03:56:40.263435884 +0000 UTC m=+1359.508187155" Dec 06 03:56:43 crc kubenswrapper[4980]: I1206 03:56:43.270501 4980 generic.go:334] "Generic (PLEG): container finished" podID="fca3de68-d822-49ad-bf87-9b0fab269bbb" containerID="fb73adf57f7fa906fd21886e3d9c5da1211cee117c1e8b663adb5a31071287d8" exitCode=0 Dec 06 03:56:43 crc kubenswrapper[4980]: I1206 03:56:43.270747 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-sync-5lft5" event={"ID":"fca3de68-d822-49ad-bf87-9b0fab269bbb","Type":"ContainerDied","Data":"fb73adf57f7fa906fd21886e3d9c5da1211cee117c1e8b663adb5a31071287d8"} Dec 06 03:56:44 crc kubenswrapper[4980]: I1206 03:56:44.588723 4980 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-sync-5lft5" Dec 06 03:56:44 crc kubenswrapper[4980]: I1206 03:56:44.777021 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fca3de68-d822-49ad-bf87-9b0fab269bbb-config-data\") pod \"fca3de68-d822-49ad-bf87-9b0fab269bbb\" (UID: \"fca3de68-d822-49ad-bf87-9b0fab269bbb\") " Dec 06 03:56:44 crc kubenswrapper[4980]: I1206 03:56:44.777127 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fca3de68-d822-49ad-bf87-9b0fab269bbb-combined-ca-bundle\") pod \"fca3de68-d822-49ad-bf87-9b0fab269bbb\" (UID: \"fca3de68-d822-49ad-bf87-9b0fab269bbb\") " Dec 06 03:56:44 crc kubenswrapper[4980]: I1206 03:56:44.777199 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/fca3de68-d822-49ad-bf87-9b0fab269bbb-db-sync-config-data\") pod \"fca3de68-d822-49ad-bf87-9b0fab269bbb\" (UID: \"fca3de68-d822-49ad-bf87-9b0fab269bbb\") " Dec 06 03:56:44 crc kubenswrapper[4980]: I1206 03:56:44.777257 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hvfng\" (UniqueName: \"kubernetes.io/projected/fca3de68-d822-49ad-bf87-9b0fab269bbb-kube-api-access-hvfng\") pod \"fca3de68-d822-49ad-bf87-9b0fab269bbb\" (UID: \"fca3de68-d822-49ad-bf87-9b0fab269bbb\") " Dec 06 03:56:44 crc kubenswrapper[4980]: I1206 03:56:44.782920 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fca3de68-d822-49ad-bf87-9b0fab269bbb-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "fca3de68-d822-49ad-bf87-9b0fab269bbb" (UID: "fca3de68-d822-49ad-bf87-9b0fab269bbb"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 03:56:44 crc kubenswrapper[4980]: I1206 03:56:44.782992 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fca3de68-d822-49ad-bf87-9b0fab269bbb-kube-api-access-hvfng" (OuterVolumeSpecName: "kube-api-access-hvfng") pod "fca3de68-d822-49ad-bf87-9b0fab269bbb" (UID: "fca3de68-d822-49ad-bf87-9b0fab269bbb"). InnerVolumeSpecName "kube-api-access-hvfng". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 03:56:44 crc kubenswrapper[4980]: I1206 03:56:44.879444 4980 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hvfng\" (UniqueName: \"kubernetes.io/projected/fca3de68-d822-49ad-bf87-9b0fab269bbb-kube-api-access-hvfng\") on node \"crc\" DevicePath \"\"" Dec 06 03:56:44 crc kubenswrapper[4980]: I1206 03:56:44.879900 4980 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/fca3de68-d822-49ad-bf87-9b0fab269bbb-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 03:56:45 crc kubenswrapper[4980]: I1206 03:56:45.290150 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-sync-5lft5" event={"ID":"fca3de68-d822-49ad-bf87-9b0fab269bbb","Type":"ContainerDied","Data":"4a6f918015ca3b061c127fdf679308be47f35733e60b4651a4bbea0d5c492b2d"} Dec 06 03:56:45 crc kubenswrapper[4980]: I1206 03:56:45.290198 4980 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4a6f918015ca3b061c127fdf679308be47f35733e60b4651a4bbea0d5c492b2d" Dec 06 03:56:45 crc kubenswrapper[4980]: I1206 03:56:45.290237 4980 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-sync-5lft5" Dec 06 03:56:45 crc kubenswrapper[4980]: I1206 03:56:45.718103 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fca3de68-d822-49ad-bf87-9b0fab269bbb-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "fca3de68-d822-49ad-bf87-9b0fab269bbb" (UID: "fca3de68-d822-49ad-bf87-9b0fab269bbb"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 03:56:45 crc kubenswrapper[4980]: I1206 03:56:45.732810 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fca3de68-d822-49ad-bf87-9b0fab269bbb-config-data" (OuterVolumeSpecName: "config-data") pod "fca3de68-d822-49ad-bf87-9b0fab269bbb" (UID: "fca3de68-d822-49ad-bf87-9b0fab269bbb"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 03:56:45 crc kubenswrapper[4980]: I1206 03:56:45.808731 4980 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fca3de68-d822-49ad-bf87-9b0fab269bbb-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 03:56:45 crc kubenswrapper[4980]: I1206 03:56:45.808763 4980 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fca3de68-d822-49ad-bf87-9b0fab269bbb-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 03:56:45 crc kubenswrapper[4980]: I1206 03:56:45.882568 4980 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Dec 06 03:56:45 crc kubenswrapper[4980]: E1206 03:56:45.882851 4980 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fca3de68-d822-49ad-bf87-9b0fab269bbb" containerName="glance-db-sync" Dec 06 03:56:45 crc kubenswrapper[4980]: I1206 03:56:45.882869 4980 state_mem.go:107] "Deleted CPUSet assignment" podUID="fca3de68-d822-49ad-bf87-9b0fab269bbb" containerName="glance-db-sync" Dec 06 03:56:45 crc kubenswrapper[4980]: I1206 03:56:45.882993 4980 memory_manager.go:354] "RemoveStaleState removing state" podUID="fca3de68-d822-49ad-bf87-9b0fab269bbb" containerName="glance-db-sync" Dec 06 03:56:45 crc kubenswrapper[4980]: I1206 03:56:45.883715 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-0" Dec 06 03:56:45 crc kubenswrapper[4980]: I1206 03:56:45.886257 4980 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-scripts" Dec 06 03:56:45 crc kubenswrapper[4980]: I1206 03:56:45.887174 4980 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-default-single-config-data" Dec 06 03:56:45 crc kubenswrapper[4980]: I1206 03:56:45.887375 4980 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"cert-glance-default-internal-svc" Dec 06 03:56:45 crc kubenswrapper[4980]: I1206 03:56:45.887576 4980 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"cert-glance-default-public-svc" Dec 06 03:56:45 crc kubenswrapper[4980]: I1206 03:56:45.942597 4980 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Dec 06 03:56:45 crc kubenswrapper[4980]: I1206 03:56:45.949014 4980 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Dec 06 03:56:45 crc kubenswrapper[4980]: E1206 03:56:45.949506 4980 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[combined-ca-bundle config-data glance httpd-run internal-tls-certs kube-api-access-l2t98 logs public-tls-certs scripts], unattached volumes=[], failed to process volumes=[]: context canceled" pod="glance-kuttl-tests/glance-default-single-0" podUID="be4510cc-5afd-4456-9613-027bee349d21" Dec 06 03:56:46 crc kubenswrapper[4980]: I1206 03:56:46.012475 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l2t98\" (UniqueName: \"kubernetes.io/projected/be4510cc-5afd-4456-9613-027bee349d21-kube-api-access-l2t98\") pod \"glance-default-single-0\" (UID: \"be4510cc-5afd-4456-9613-027bee349d21\") " pod="glance-kuttl-tests/glance-default-single-0" Dec 06 03:56:46 crc kubenswrapper[4980]: I1206 03:56:46.012585 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/be4510cc-5afd-4456-9613-027bee349d21-scripts\") pod \"glance-default-single-0\" (UID: \"be4510cc-5afd-4456-9613-027bee349d21\") " pod="glance-kuttl-tests/glance-default-single-0" Dec 06 03:56:46 crc kubenswrapper[4980]: I1206 03:56:46.012668 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/be4510cc-5afd-4456-9613-027bee349d21-logs\") pod \"glance-default-single-0\" (UID: \"be4510cc-5afd-4456-9613-027bee349d21\") " pod="glance-kuttl-tests/glance-default-single-0" Dec 06 03:56:46 crc kubenswrapper[4980]: I1206 03:56:46.012707 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/be4510cc-5afd-4456-9613-027bee349d21-config-data\") pod \"glance-default-single-0\" (UID: \"be4510cc-5afd-4456-9613-027bee349d21\") " pod="glance-kuttl-tests/glance-default-single-0" Dec 06 03:56:46 crc kubenswrapper[4980]: I1206 03:56:46.012736 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/be4510cc-5afd-4456-9613-027bee349d21-httpd-run\") pod \"glance-default-single-0\" (UID: \"be4510cc-5afd-4456-9613-027bee349d21\") " pod="glance-kuttl-tests/glance-default-single-0" Dec 06 03:56:46 crc kubenswrapper[4980]: I1206 03:56:46.012775 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/be4510cc-5afd-4456-9613-027bee349d21-public-tls-certs\") pod \"glance-default-single-0\" (UID: \"be4510cc-5afd-4456-9613-027bee349d21\") " pod="glance-kuttl-tests/glance-default-single-0" Dec 06 03:56:46 crc kubenswrapper[4980]: I1206 03:56:46.012821 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-single-0\" (UID: \"be4510cc-5afd-4456-9613-027bee349d21\") " pod="glance-kuttl-tests/glance-default-single-0" Dec 06 03:56:46 crc kubenswrapper[4980]: I1206 03:56:46.012993 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/be4510cc-5afd-4456-9613-027bee349d21-internal-tls-certs\") pod \"glance-default-single-0\" (UID: \"be4510cc-5afd-4456-9613-027bee349d21\") " pod="glance-kuttl-tests/glance-default-single-0" Dec 06 03:56:46 crc kubenswrapper[4980]: I1206 03:56:46.013166 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/be4510cc-5afd-4456-9613-027bee349d21-combined-ca-bundle\") pod \"glance-default-single-0\" (UID: \"be4510cc-5afd-4456-9613-027bee349d21\") " pod="glance-kuttl-tests/glance-default-single-0" Dec 06 03:56:46 crc kubenswrapper[4980]: I1206 03:56:46.114668 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/be4510cc-5afd-4456-9613-027bee349d21-public-tls-certs\") pod \"glance-default-single-0\" (UID: \"be4510cc-5afd-4456-9613-027bee349d21\") " pod="glance-kuttl-tests/glance-default-single-0" Dec 06 03:56:46 crc kubenswrapper[4980]: I1206 03:56:46.114743 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-single-0\" (UID: \"be4510cc-5afd-4456-9613-027bee349d21\") " pod="glance-kuttl-tests/glance-default-single-0" Dec 06 03:56:46 crc kubenswrapper[4980]: I1206 03:56:46.114775 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/be4510cc-5afd-4456-9613-027bee349d21-internal-tls-certs\") pod \"glance-default-single-0\" (UID: \"be4510cc-5afd-4456-9613-027bee349d21\") " pod="glance-kuttl-tests/glance-default-single-0" Dec 06 03:56:46 crc kubenswrapper[4980]: I1206 03:56:46.114820 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/be4510cc-5afd-4456-9613-027bee349d21-combined-ca-bundle\") pod \"glance-default-single-0\" (UID: \"be4510cc-5afd-4456-9613-027bee349d21\") " pod="glance-kuttl-tests/glance-default-single-0" Dec 06 03:56:46 crc kubenswrapper[4980]: I1206 03:56:46.114883 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l2t98\" (UniqueName: \"kubernetes.io/projected/be4510cc-5afd-4456-9613-027bee349d21-kube-api-access-l2t98\") pod \"glance-default-single-0\" (UID: \"be4510cc-5afd-4456-9613-027bee349d21\") " pod="glance-kuttl-tests/glance-default-single-0" Dec 06 03:56:46 crc kubenswrapper[4980]: I1206 03:56:46.114906 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/be4510cc-5afd-4456-9613-027bee349d21-scripts\") pod \"glance-default-single-0\" (UID: \"be4510cc-5afd-4456-9613-027bee349d21\") " pod="glance-kuttl-tests/glance-default-single-0" Dec 06 03:56:46 crc kubenswrapper[4980]: I1206 03:56:46.114931 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/be4510cc-5afd-4456-9613-027bee349d21-logs\") pod \"glance-default-single-0\" (UID: \"be4510cc-5afd-4456-9613-027bee349d21\") " pod="glance-kuttl-tests/glance-default-single-0" Dec 06 03:56:46 crc kubenswrapper[4980]: I1206 03:56:46.114958 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/be4510cc-5afd-4456-9613-027bee349d21-config-data\") pod \"glance-default-single-0\" (UID: \"be4510cc-5afd-4456-9613-027bee349d21\") " pod="glance-kuttl-tests/glance-default-single-0" Dec 06 03:56:46 crc kubenswrapper[4980]: I1206 03:56:46.114978 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/be4510cc-5afd-4456-9613-027bee349d21-httpd-run\") pod \"glance-default-single-0\" (UID: \"be4510cc-5afd-4456-9613-027bee349d21\") " pod="glance-kuttl-tests/glance-default-single-0" Dec 06 03:56:46 crc kubenswrapper[4980]: I1206 03:56:46.115585 4980 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-single-0\" (UID: \"be4510cc-5afd-4456-9613-027bee349d21\") device mount path \"/mnt/openstack/pv09\"" pod="glance-kuttl-tests/glance-default-single-0" Dec 06 03:56:46 crc kubenswrapper[4980]: I1206 03:56:46.116201 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/be4510cc-5afd-4456-9613-027bee349d21-logs\") pod \"glance-default-single-0\" (UID: \"be4510cc-5afd-4456-9613-027bee349d21\") " pod="glance-kuttl-tests/glance-default-single-0" Dec 06 03:56:46 crc kubenswrapper[4980]: I1206 03:56:46.116640 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/be4510cc-5afd-4456-9613-027bee349d21-httpd-run\") pod \"glance-default-single-0\" (UID: \"be4510cc-5afd-4456-9613-027bee349d21\") " pod="glance-kuttl-tests/glance-default-single-0" Dec 06 03:56:46 crc kubenswrapper[4980]: I1206 03:56:46.119444 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/be4510cc-5afd-4456-9613-027bee349d21-internal-tls-certs\") pod \"glance-default-single-0\" (UID: \"be4510cc-5afd-4456-9613-027bee349d21\") " pod="glance-kuttl-tests/glance-default-single-0" Dec 06 03:56:46 crc kubenswrapper[4980]: I1206 03:56:46.120232 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/be4510cc-5afd-4456-9613-027bee349d21-scripts\") pod \"glance-default-single-0\" (UID: \"be4510cc-5afd-4456-9613-027bee349d21\") " pod="glance-kuttl-tests/glance-default-single-0" Dec 06 03:56:46 crc kubenswrapper[4980]: I1206 03:56:46.120685 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/be4510cc-5afd-4456-9613-027bee349d21-public-tls-certs\") pod \"glance-default-single-0\" (UID: \"be4510cc-5afd-4456-9613-027bee349d21\") " pod="glance-kuttl-tests/glance-default-single-0" Dec 06 03:56:46 crc kubenswrapper[4980]: I1206 03:56:46.121014 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/be4510cc-5afd-4456-9613-027bee349d21-combined-ca-bundle\") pod \"glance-default-single-0\" (UID: \"be4510cc-5afd-4456-9613-027bee349d21\") " pod="glance-kuttl-tests/glance-default-single-0" Dec 06 03:56:46 crc kubenswrapper[4980]: I1206 03:56:46.121884 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/be4510cc-5afd-4456-9613-027bee349d21-config-data\") pod \"glance-default-single-0\" (UID: \"be4510cc-5afd-4456-9613-027bee349d21\") " pod="glance-kuttl-tests/glance-default-single-0" Dec 06 03:56:46 crc kubenswrapper[4980]: I1206 03:56:46.150770 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l2t98\" (UniqueName: \"kubernetes.io/projected/be4510cc-5afd-4456-9613-027bee349d21-kube-api-access-l2t98\") pod \"glance-default-single-0\" (UID: \"be4510cc-5afd-4456-9613-027bee349d21\") " pod="glance-kuttl-tests/glance-default-single-0" Dec 06 03:56:46 crc kubenswrapper[4980]: I1206 03:56:46.153974 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-single-0\" (UID: \"be4510cc-5afd-4456-9613-027bee349d21\") " pod="glance-kuttl-tests/glance-default-single-0" Dec 06 03:56:46 crc kubenswrapper[4980]: I1206 03:56:46.296678 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-0" Dec 06 03:56:46 crc kubenswrapper[4980]: I1206 03:56:46.305424 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-0" Dec 06 03:56:46 crc kubenswrapper[4980]: I1206 03:56:46.418506 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/be4510cc-5afd-4456-9613-027bee349d21-config-data\") pod \"be4510cc-5afd-4456-9613-027bee349d21\" (UID: \"be4510cc-5afd-4456-9613-027bee349d21\") " Dec 06 03:56:46 crc kubenswrapper[4980]: I1206 03:56:46.418618 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/be4510cc-5afd-4456-9613-027bee349d21-logs\") pod \"be4510cc-5afd-4456-9613-027bee349d21\" (UID: \"be4510cc-5afd-4456-9613-027bee349d21\") " Dec 06 03:56:46 crc kubenswrapper[4980]: I1206 03:56:46.418640 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/be4510cc-5afd-4456-9613-027bee349d21-scripts\") pod \"be4510cc-5afd-4456-9613-027bee349d21\" (UID: \"be4510cc-5afd-4456-9613-027bee349d21\") " Dec 06 03:56:46 crc kubenswrapper[4980]: I1206 03:56:46.418677 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/be4510cc-5afd-4456-9613-027bee349d21-public-tls-certs\") pod \"be4510cc-5afd-4456-9613-027bee349d21\" (UID: \"be4510cc-5afd-4456-9613-027bee349d21\") " Dec 06 03:56:46 crc kubenswrapper[4980]: I1206 03:56:46.418701 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/be4510cc-5afd-4456-9613-027bee349d21-combined-ca-bundle\") pod \"be4510cc-5afd-4456-9613-027bee349d21\" (UID: \"be4510cc-5afd-4456-9613-027bee349d21\") " Dec 06 03:56:46 crc kubenswrapper[4980]: I1206 03:56:46.418725 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/be4510cc-5afd-4456-9613-027bee349d21-internal-tls-certs\") pod \"be4510cc-5afd-4456-9613-027bee349d21\" (UID: \"be4510cc-5afd-4456-9613-027bee349d21\") " Dec 06 03:56:46 crc kubenswrapper[4980]: I1206 03:56:46.418798 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l2t98\" (UniqueName: \"kubernetes.io/projected/be4510cc-5afd-4456-9613-027bee349d21-kube-api-access-l2t98\") pod \"be4510cc-5afd-4456-9613-027bee349d21\" (UID: \"be4510cc-5afd-4456-9613-027bee349d21\") " Dec 06 03:56:46 crc kubenswrapper[4980]: I1206 03:56:46.418813 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"be4510cc-5afd-4456-9613-027bee349d21\" (UID: \"be4510cc-5afd-4456-9613-027bee349d21\") " Dec 06 03:56:46 crc kubenswrapper[4980]: I1206 03:56:46.418893 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/be4510cc-5afd-4456-9613-027bee349d21-httpd-run\") pod \"be4510cc-5afd-4456-9613-027bee349d21\" (UID: \"be4510cc-5afd-4456-9613-027bee349d21\") " Dec 06 03:56:46 crc kubenswrapper[4980]: I1206 03:56:46.419489 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/be4510cc-5afd-4456-9613-027bee349d21-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "be4510cc-5afd-4456-9613-027bee349d21" (UID: "be4510cc-5afd-4456-9613-027bee349d21"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 03:56:46 crc kubenswrapper[4980]: I1206 03:56:46.420159 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/be4510cc-5afd-4456-9613-027bee349d21-logs" (OuterVolumeSpecName: "logs") pod "be4510cc-5afd-4456-9613-027bee349d21" (UID: "be4510cc-5afd-4456-9613-027bee349d21"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 03:56:46 crc kubenswrapper[4980]: I1206 03:56:46.423500 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/be4510cc-5afd-4456-9613-027bee349d21-config-data" (OuterVolumeSpecName: "config-data") pod "be4510cc-5afd-4456-9613-027bee349d21" (UID: "be4510cc-5afd-4456-9613-027bee349d21"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 03:56:46 crc kubenswrapper[4980]: I1206 03:56:46.424287 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/be4510cc-5afd-4456-9613-027bee349d21-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "be4510cc-5afd-4456-9613-027bee349d21" (UID: "be4510cc-5afd-4456-9613-027bee349d21"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 03:56:46 crc kubenswrapper[4980]: I1206 03:56:46.424661 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/be4510cc-5afd-4456-9613-027bee349d21-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "be4510cc-5afd-4456-9613-027bee349d21" (UID: "be4510cc-5afd-4456-9613-027bee349d21"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 03:56:46 crc kubenswrapper[4980]: I1206 03:56:46.426656 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/be4510cc-5afd-4456-9613-027bee349d21-scripts" (OuterVolumeSpecName: "scripts") pod "be4510cc-5afd-4456-9613-027bee349d21" (UID: "be4510cc-5afd-4456-9613-027bee349d21"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 03:56:46 crc kubenswrapper[4980]: I1206 03:56:46.427161 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/be4510cc-5afd-4456-9613-027bee349d21-kube-api-access-l2t98" (OuterVolumeSpecName: "kube-api-access-l2t98") pod "be4510cc-5afd-4456-9613-027bee349d21" (UID: "be4510cc-5afd-4456-9613-027bee349d21"). InnerVolumeSpecName "kube-api-access-l2t98". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 03:56:46 crc kubenswrapper[4980]: I1206 03:56:46.437570 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage09-crc" (OuterVolumeSpecName: "glance") pod "be4510cc-5afd-4456-9613-027bee349d21" (UID: "be4510cc-5afd-4456-9613-027bee349d21"). InnerVolumeSpecName "local-storage09-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 06 03:56:46 crc kubenswrapper[4980]: I1206 03:56:46.442573 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/be4510cc-5afd-4456-9613-027bee349d21-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "be4510cc-5afd-4456-9613-027bee349d21" (UID: "be4510cc-5afd-4456-9613-027bee349d21"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 03:56:46 crc kubenswrapper[4980]: I1206 03:56:46.520380 4980 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l2t98\" (UniqueName: \"kubernetes.io/projected/be4510cc-5afd-4456-9613-027bee349d21-kube-api-access-l2t98\") on node \"crc\" DevicePath \"\"" Dec 06 03:56:46 crc kubenswrapper[4980]: I1206 03:56:46.520441 4980 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") on node \"crc\" " Dec 06 03:56:46 crc kubenswrapper[4980]: I1206 03:56:46.520451 4980 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/be4510cc-5afd-4456-9613-027bee349d21-httpd-run\") on node \"crc\" DevicePath \"\"" Dec 06 03:56:46 crc kubenswrapper[4980]: I1206 03:56:46.520460 4980 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/be4510cc-5afd-4456-9613-027bee349d21-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 03:56:46 crc kubenswrapper[4980]: I1206 03:56:46.520468 4980 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/be4510cc-5afd-4456-9613-027bee349d21-logs\") on node \"crc\" DevicePath \"\"" Dec 06 03:56:46 crc kubenswrapper[4980]: I1206 03:56:46.520477 4980 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/be4510cc-5afd-4456-9613-027bee349d21-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 03:56:46 crc kubenswrapper[4980]: I1206 03:56:46.520485 4980 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/be4510cc-5afd-4456-9613-027bee349d21-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 03:56:46 crc kubenswrapper[4980]: I1206 03:56:46.520493 4980 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/be4510cc-5afd-4456-9613-027bee349d21-public-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 06 03:56:46 crc kubenswrapper[4980]: I1206 03:56:46.520500 4980 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/be4510cc-5afd-4456-9613-027bee349d21-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 06 03:56:46 crc kubenswrapper[4980]: I1206 03:56:46.534169 4980 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage09-crc" (UniqueName: "kubernetes.io/local-volume/local-storage09-crc") on node "crc" Dec 06 03:56:46 crc kubenswrapper[4980]: I1206 03:56:46.621866 4980 reconciler_common.go:293] "Volume detached for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") on node \"crc\" DevicePath \"\"" Dec 06 03:56:47 crc kubenswrapper[4980]: I1206 03:56:47.302887 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-0" Dec 06 03:56:47 crc kubenswrapper[4980]: I1206 03:56:47.343434 4980 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Dec 06 03:56:47 crc kubenswrapper[4980]: I1206 03:56:47.353722 4980 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Dec 06 03:56:47 crc kubenswrapper[4980]: I1206 03:56:47.398799 4980 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Dec 06 03:56:47 crc kubenswrapper[4980]: I1206 03:56:47.400018 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-0" Dec 06 03:56:47 crc kubenswrapper[4980]: I1206 03:56:47.407771 4980 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-scripts" Dec 06 03:56:47 crc kubenswrapper[4980]: I1206 03:56:47.407778 4980 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"combined-ca-bundle" Dec 06 03:56:47 crc kubenswrapper[4980]: I1206 03:56:47.408198 4980 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-glance-dockercfg-kjjjr" Dec 06 03:56:47 crc kubenswrapper[4980]: I1206 03:56:47.408348 4980 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"cert-glance-default-public-svc" Dec 06 03:56:47 crc kubenswrapper[4980]: I1206 03:56:47.408601 4980 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-default-single-config-data" Dec 06 03:56:47 crc kubenswrapper[4980]: I1206 03:56:47.409228 4980 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"cert-glance-default-internal-svc" Dec 06 03:56:47 crc kubenswrapper[4980]: I1206 03:56:47.414709 4980 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Dec 06 03:56:47 crc kubenswrapper[4980]: I1206 03:56:47.554660 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nmsg2\" (UniqueName: \"kubernetes.io/projected/9f0d1967-5bde-442f-9077-f7871b23df63-kube-api-access-nmsg2\") pod \"glance-default-single-0\" (UID: \"9f0d1967-5bde-442f-9077-f7871b23df63\") " pod="glance-kuttl-tests/glance-default-single-0" Dec 06 03:56:47 crc kubenswrapper[4980]: I1206 03:56:47.554728 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9f0d1967-5bde-442f-9077-f7871b23df63-logs\") pod \"glance-default-single-0\" (UID: \"9f0d1967-5bde-442f-9077-f7871b23df63\") " pod="glance-kuttl-tests/glance-default-single-0" Dec 06 03:56:47 crc kubenswrapper[4980]: I1206 03:56:47.554775 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/9f0d1967-5bde-442f-9077-f7871b23df63-internal-tls-certs\") pod \"glance-default-single-0\" (UID: \"9f0d1967-5bde-442f-9077-f7871b23df63\") " pod="glance-kuttl-tests/glance-default-single-0" Dec 06 03:56:47 crc kubenswrapper[4980]: I1206 03:56:47.554802 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9f0d1967-5bde-442f-9077-f7871b23df63-combined-ca-bundle\") pod \"glance-default-single-0\" (UID: \"9f0d1967-5bde-442f-9077-f7871b23df63\") " pod="glance-kuttl-tests/glance-default-single-0" Dec 06 03:56:47 crc kubenswrapper[4980]: I1206 03:56:47.554862 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9f0d1967-5bde-442f-9077-f7871b23df63-scripts\") pod \"glance-default-single-0\" (UID: \"9f0d1967-5bde-442f-9077-f7871b23df63\") " pod="glance-kuttl-tests/glance-default-single-0" Dec 06 03:56:47 crc kubenswrapper[4980]: I1206 03:56:47.554901 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/9f0d1967-5bde-442f-9077-f7871b23df63-public-tls-certs\") pod \"glance-default-single-0\" (UID: \"9f0d1967-5bde-442f-9077-f7871b23df63\") " pod="glance-kuttl-tests/glance-default-single-0" Dec 06 03:56:47 crc kubenswrapper[4980]: I1206 03:56:47.554929 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/9f0d1967-5bde-442f-9077-f7871b23df63-httpd-run\") pod \"glance-default-single-0\" (UID: \"9f0d1967-5bde-442f-9077-f7871b23df63\") " pod="glance-kuttl-tests/glance-default-single-0" Dec 06 03:56:47 crc kubenswrapper[4980]: I1206 03:56:47.554953 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-single-0\" (UID: \"9f0d1967-5bde-442f-9077-f7871b23df63\") " pod="glance-kuttl-tests/glance-default-single-0" Dec 06 03:56:47 crc kubenswrapper[4980]: I1206 03:56:47.554973 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9f0d1967-5bde-442f-9077-f7871b23df63-config-data\") pod \"glance-default-single-0\" (UID: \"9f0d1967-5bde-442f-9077-f7871b23df63\") " pod="glance-kuttl-tests/glance-default-single-0" Dec 06 03:56:47 crc kubenswrapper[4980]: I1206 03:56:47.656257 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9f0d1967-5bde-442f-9077-f7871b23df63-scripts\") pod \"glance-default-single-0\" (UID: \"9f0d1967-5bde-442f-9077-f7871b23df63\") " pod="glance-kuttl-tests/glance-default-single-0" Dec 06 03:56:47 crc kubenswrapper[4980]: I1206 03:56:47.656334 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/9f0d1967-5bde-442f-9077-f7871b23df63-public-tls-certs\") pod \"glance-default-single-0\" (UID: \"9f0d1967-5bde-442f-9077-f7871b23df63\") " pod="glance-kuttl-tests/glance-default-single-0" Dec 06 03:56:47 crc kubenswrapper[4980]: I1206 03:56:47.656368 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/9f0d1967-5bde-442f-9077-f7871b23df63-httpd-run\") pod \"glance-default-single-0\" (UID: \"9f0d1967-5bde-442f-9077-f7871b23df63\") " pod="glance-kuttl-tests/glance-default-single-0" Dec 06 03:56:47 crc kubenswrapper[4980]: I1206 03:56:47.656389 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9f0d1967-5bde-442f-9077-f7871b23df63-config-data\") pod \"glance-default-single-0\" (UID: \"9f0d1967-5bde-442f-9077-f7871b23df63\") " pod="glance-kuttl-tests/glance-default-single-0" Dec 06 03:56:47 crc kubenswrapper[4980]: I1206 03:56:47.656410 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-single-0\" (UID: \"9f0d1967-5bde-442f-9077-f7871b23df63\") " pod="glance-kuttl-tests/glance-default-single-0" Dec 06 03:56:47 crc kubenswrapper[4980]: I1206 03:56:47.656464 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nmsg2\" (UniqueName: \"kubernetes.io/projected/9f0d1967-5bde-442f-9077-f7871b23df63-kube-api-access-nmsg2\") pod \"glance-default-single-0\" (UID: \"9f0d1967-5bde-442f-9077-f7871b23df63\") " pod="glance-kuttl-tests/glance-default-single-0" Dec 06 03:56:47 crc kubenswrapper[4980]: I1206 03:56:47.656502 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9f0d1967-5bde-442f-9077-f7871b23df63-logs\") pod \"glance-default-single-0\" (UID: \"9f0d1967-5bde-442f-9077-f7871b23df63\") " pod="glance-kuttl-tests/glance-default-single-0" Dec 06 03:56:47 crc kubenswrapper[4980]: I1206 03:56:47.656659 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/9f0d1967-5bde-442f-9077-f7871b23df63-internal-tls-certs\") pod \"glance-default-single-0\" (UID: \"9f0d1967-5bde-442f-9077-f7871b23df63\") " pod="glance-kuttl-tests/glance-default-single-0" Dec 06 03:56:47 crc kubenswrapper[4980]: I1206 03:56:47.656686 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9f0d1967-5bde-442f-9077-f7871b23df63-combined-ca-bundle\") pod \"glance-default-single-0\" (UID: \"9f0d1967-5bde-442f-9077-f7871b23df63\") " pod="glance-kuttl-tests/glance-default-single-0" Dec 06 03:56:47 crc kubenswrapper[4980]: I1206 03:56:47.656723 4980 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-single-0\" (UID: \"9f0d1967-5bde-442f-9077-f7871b23df63\") device mount path \"/mnt/openstack/pv09\"" pod="glance-kuttl-tests/glance-default-single-0" Dec 06 03:56:47 crc kubenswrapper[4980]: I1206 03:56:47.656865 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/9f0d1967-5bde-442f-9077-f7871b23df63-httpd-run\") pod \"glance-default-single-0\" (UID: \"9f0d1967-5bde-442f-9077-f7871b23df63\") " pod="glance-kuttl-tests/glance-default-single-0" Dec 06 03:56:47 crc kubenswrapper[4980]: I1206 03:56:47.657336 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9f0d1967-5bde-442f-9077-f7871b23df63-logs\") pod \"glance-default-single-0\" (UID: \"9f0d1967-5bde-442f-9077-f7871b23df63\") " pod="glance-kuttl-tests/glance-default-single-0" Dec 06 03:56:47 crc kubenswrapper[4980]: I1206 03:56:47.661936 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/9f0d1967-5bde-442f-9077-f7871b23df63-internal-tls-certs\") pod \"glance-default-single-0\" (UID: \"9f0d1967-5bde-442f-9077-f7871b23df63\") " pod="glance-kuttl-tests/glance-default-single-0" Dec 06 03:56:47 crc kubenswrapper[4980]: I1206 03:56:47.662642 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/9f0d1967-5bde-442f-9077-f7871b23df63-public-tls-certs\") pod \"glance-default-single-0\" (UID: \"9f0d1967-5bde-442f-9077-f7871b23df63\") " pod="glance-kuttl-tests/glance-default-single-0" Dec 06 03:56:47 crc kubenswrapper[4980]: I1206 03:56:47.669328 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9f0d1967-5bde-442f-9077-f7871b23df63-scripts\") pod \"glance-default-single-0\" (UID: \"9f0d1967-5bde-442f-9077-f7871b23df63\") " pod="glance-kuttl-tests/glance-default-single-0" Dec 06 03:56:47 crc kubenswrapper[4980]: I1206 03:56:47.671363 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9f0d1967-5bde-442f-9077-f7871b23df63-combined-ca-bundle\") pod \"glance-default-single-0\" (UID: \"9f0d1967-5bde-442f-9077-f7871b23df63\") " pod="glance-kuttl-tests/glance-default-single-0" Dec 06 03:56:47 crc kubenswrapper[4980]: I1206 03:56:47.671786 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9f0d1967-5bde-442f-9077-f7871b23df63-config-data\") pod \"glance-default-single-0\" (UID: \"9f0d1967-5bde-442f-9077-f7871b23df63\") " pod="glance-kuttl-tests/glance-default-single-0" Dec 06 03:56:47 crc kubenswrapper[4980]: I1206 03:56:47.679194 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-single-0\" (UID: \"9f0d1967-5bde-442f-9077-f7871b23df63\") " pod="glance-kuttl-tests/glance-default-single-0" Dec 06 03:56:47 crc kubenswrapper[4980]: I1206 03:56:47.687333 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nmsg2\" (UniqueName: \"kubernetes.io/projected/9f0d1967-5bde-442f-9077-f7871b23df63-kube-api-access-nmsg2\") pod \"glance-default-single-0\" (UID: \"9f0d1967-5bde-442f-9077-f7871b23df63\") " pod="glance-kuttl-tests/glance-default-single-0" Dec 06 03:56:47 crc kubenswrapper[4980]: I1206 03:56:47.728251 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-0" Dec 06 03:56:47 crc kubenswrapper[4980]: I1206 03:56:47.950470 4980 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Dec 06 03:56:48 crc kubenswrapper[4980]: I1206 03:56:48.314675 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"9f0d1967-5bde-442f-9077-f7871b23df63","Type":"ContainerStarted","Data":"0cb082b3657cc6f0e6bc566a3c1bc7de7dcd8314718289e29c88b52bd2525622"} Dec 06 03:56:48 crc kubenswrapper[4980]: I1206 03:56:48.741767 4980 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-nzvgz"] Dec 06 03:56:48 crc kubenswrapper[4980]: I1206 03:56:48.759314 4980 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-nzvgz"] Dec 06 03:56:48 crc kubenswrapper[4980]: I1206 03:56:48.762936 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-nzvgz" Dec 06 03:56:48 crc kubenswrapper[4980]: I1206 03:56:48.874747 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4dtg5\" (UniqueName: \"kubernetes.io/projected/d45ee895-b91f-4156-a591-94d6d750a5b2-kube-api-access-4dtg5\") pod \"redhat-operators-nzvgz\" (UID: \"d45ee895-b91f-4156-a591-94d6d750a5b2\") " pod="openshift-marketplace/redhat-operators-nzvgz" Dec 06 03:56:48 crc kubenswrapper[4980]: I1206 03:56:48.874863 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d45ee895-b91f-4156-a591-94d6d750a5b2-catalog-content\") pod \"redhat-operators-nzvgz\" (UID: \"d45ee895-b91f-4156-a591-94d6d750a5b2\") " pod="openshift-marketplace/redhat-operators-nzvgz" Dec 06 03:56:48 crc kubenswrapper[4980]: I1206 03:56:48.874905 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d45ee895-b91f-4156-a591-94d6d750a5b2-utilities\") pod \"redhat-operators-nzvgz\" (UID: \"d45ee895-b91f-4156-a591-94d6d750a5b2\") " pod="openshift-marketplace/redhat-operators-nzvgz" Dec 06 03:56:48 crc kubenswrapper[4980]: I1206 03:56:48.976779 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4dtg5\" (UniqueName: \"kubernetes.io/projected/d45ee895-b91f-4156-a591-94d6d750a5b2-kube-api-access-4dtg5\") pod \"redhat-operators-nzvgz\" (UID: \"d45ee895-b91f-4156-a591-94d6d750a5b2\") " pod="openshift-marketplace/redhat-operators-nzvgz" Dec 06 03:56:48 crc kubenswrapper[4980]: I1206 03:56:48.976860 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d45ee895-b91f-4156-a591-94d6d750a5b2-catalog-content\") pod \"redhat-operators-nzvgz\" (UID: \"d45ee895-b91f-4156-a591-94d6d750a5b2\") " pod="openshift-marketplace/redhat-operators-nzvgz" Dec 06 03:56:48 crc kubenswrapper[4980]: I1206 03:56:48.976900 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d45ee895-b91f-4156-a591-94d6d750a5b2-utilities\") pod \"redhat-operators-nzvgz\" (UID: \"d45ee895-b91f-4156-a591-94d6d750a5b2\") " pod="openshift-marketplace/redhat-operators-nzvgz" Dec 06 03:56:48 crc kubenswrapper[4980]: I1206 03:56:48.977450 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d45ee895-b91f-4156-a591-94d6d750a5b2-utilities\") pod \"redhat-operators-nzvgz\" (UID: \"d45ee895-b91f-4156-a591-94d6d750a5b2\") " pod="openshift-marketplace/redhat-operators-nzvgz" Dec 06 03:56:48 crc kubenswrapper[4980]: I1206 03:56:48.977538 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d45ee895-b91f-4156-a591-94d6d750a5b2-catalog-content\") pod \"redhat-operators-nzvgz\" (UID: \"d45ee895-b91f-4156-a591-94d6d750a5b2\") " pod="openshift-marketplace/redhat-operators-nzvgz" Dec 06 03:56:48 crc kubenswrapper[4980]: I1206 03:56:48.997063 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4dtg5\" (UniqueName: \"kubernetes.io/projected/d45ee895-b91f-4156-a591-94d6d750a5b2-kube-api-access-4dtg5\") pod \"redhat-operators-nzvgz\" (UID: \"d45ee895-b91f-4156-a591-94d6d750a5b2\") " pod="openshift-marketplace/redhat-operators-nzvgz" Dec 06 03:56:49 crc kubenswrapper[4980]: I1206 03:56:49.123273 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-nzvgz" Dec 06 03:56:49 crc kubenswrapper[4980]: I1206 03:56:49.191926 4980 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="be4510cc-5afd-4456-9613-027bee349d21" path="/var/lib/kubelet/pods/be4510cc-5afd-4456-9613-027bee349d21/volumes" Dec 06 03:56:49 crc kubenswrapper[4980]: I1206 03:56:49.324952 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"9f0d1967-5bde-442f-9077-f7871b23df63","Type":"ContainerStarted","Data":"03483b8c9a8dd9567feb893a1e8f4e2c8659385420a7a4b0c570edab3f530168"} Dec 06 03:56:49 crc kubenswrapper[4980]: I1206 03:56:49.325267 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"9f0d1967-5bde-442f-9077-f7871b23df63","Type":"ContainerStarted","Data":"84a936e4a197ceb05355a11151840749362c4751bf38242f2c3e953c7ed3e08c"} Dec 06 03:56:49 crc kubenswrapper[4980]: I1206 03:56:49.354225 4980 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/glance-default-single-0" podStartSLOduration=2.354204898 podStartE2EDuration="2.354204898s" podCreationTimestamp="2025-12-06 03:56:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 03:56:49.352675454 +0000 UTC m=+1368.597426735" watchObservedRunningTime="2025-12-06 03:56:49.354204898 +0000 UTC m=+1368.598956169" Dec 06 03:56:49 crc kubenswrapper[4980]: I1206 03:56:49.579031 4980 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-nzvgz"] Dec 06 03:56:49 crc kubenswrapper[4980]: W1206 03:56:49.580372 4980 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd45ee895_b91f_4156_a591_94d6d750a5b2.slice/crio-5ecf1a9b020dc94c0e04f3ee99debc6abfe999d6869657d24fb5195b53aa6ba9 WatchSource:0}: Error finding container 5ecf1a9b020dc94c0e04f3ee99debc6abfe999d6869657d24fb5195b53aa6ba9: Status 404 returned error can't find the container with id 5ecf1a9b020dc94c0e04f3ee99debc6abfe999d6869657d24fb5195b53aa6ba9 Dec 06 03:56:50 crc kubenswrapper[4980]: I1206 03:56:50.334071 4980 generic.go:334] "Generic (PLEG): container finished" podID="d45ee895-b91f-4156-a591-94d6d750a5b2" containerID="170f71bef6d6562f39713974265d6189fc745861b19676e72a580d32c9c5e8ce" exitCode=0 Dec 06 03:56:50 crc kubenswrapper[4980]: I1206 03:56:50.334139 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-nzvgz" event={"ID":"d45ee895-b91f-4156-a591-94d6d750a5b2","Type":"ContainerDied","Data":"170f71bef6d6562f39713974265d6189fc745861b19676e72a580d32c9c5e8ce"} Dec 06 03:56:50 crc kubenswrapper[4980]: I1206 03:56:50.334212 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-nzvgz" event={"ID":"d45ee895-b91f-4156-a591-94d6d750a5b2","Type":"ContainerStarted","Data":"5ecf1a9b020dc94c0e04f3ee99debc6abfe999d6869657d24fb5195b53aa6ba9"} Dec 06 03:56:50 crc kubenswrapper[4980]: I1206 03:56:50.336341 4980 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 06 03:56:51 crc kubenswrapper[4980]: I1206 03:56:51.356751 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-nzvgz" event={"ID":"d45ee895-b91f-4156-a591-94d6d750a5b2","Type":"ContainerStarted","Data":"df7b57e93ae628bf8385c5aec97017b17e1e1b6b4ad00ce9f71415b5afa324eb"} Dec 06 03:56:52 crc kubenswrapper[4980]: I1206 03:56:52.364680 4980 generic.go:334] "Generic (PLEG): container finished" podID="d45ee895-b91f-4156-a591-94d6d750a5b2" containerID="df7b57e93ae628bf8385c5aec97017b17e1e1b6b4ad00ce9f71415b5afa324eb" exitCode=0 Dec 06 03:56:52 crc kubenswrapper[4980]: I1206 03:56:52.364904 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-nzvgz" event={"ID":"d45ee895-b91f-4156-a591-94d6d750a5b2","Type":"ContainerDied","Data":"df7b57e93ae628bf8385c5aec97017b17e1e1b6b4ad00ce9f71415b5afa324eb"} Dec 06 03:56:53 crc kubenswrapper[4980]: I1206 03:56:53.376389 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-nzvgz" event={"ID":"d45ee895-b91f-4156-a591-94d6d750a5b2","Type":"ContainerStarted","Data":"84ecfc50e7fc781102dbe47be76e32b118fb909f0c51e2274bc301b66e9caab3"} Dec 06 03:56:53 crc kubenswrapper[4980]: I1206 03:56:53.405213 4980 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-nzvgz" podStartSLOduration=2.820587932 podStartE2EDuration="5.405188138s" podCreationTimestamp="2025-12-06 03:56:48 +0000 UTC" firstStartedPulling="2025-12-06 03:56:50.335804056 +0000 UTC m=+1369.580555327" lastFinishedPulling="2025-12-06 03:56:52.920404262 +0000 UTC m=+1372.165155533" observedRunningTime="2025-12-06 03:56:53.39621401 +0000 UTC m=+1372.640965291" watchObservedRunningTime="2025-12-06 03:56:53.405188138 +0000 UTC m=+1372.649939409" Dec 06 03:56:57 crc kubenswrapper[4980]: I1206 03:56:57.728803 4980 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-single-0" Dec 06 03:56:57 crc kubenswrapper[4980]: I1206 03:56:57.729742 4980 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-single-0" Dec 06 03:56:57 crc kubenswrapper[4980]: I1206 03:56:57.768292 4980 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-single-0" Dec 06 03:56:57 crc kubenswrapper[4980]: I1206 03:56:57.792372 4980 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-single-0" Dec 06 03:56:58 crc kubenswrapper[4980]: I1206 03:56:58.414551 4980 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-single-0" Dec 06 03:56:58 crc kubenswrapper[4980]: I1206 03:56:58.414638 4980 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-single-0" Dec 06 03:56:59 crc kubenswrapper[4980]: I1206 03:56:59.123771 4980 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-nzvgz" Dec 06 03:56:59 crc kubenswrapper[4980]: I1206 03:56:59.123859 4980 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-nzvgz" Dec 06 03:56:59 crc kubenswrapper[4980]: I1206 03:56:59.176891 4980 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-nzvgz" Dec 06 03:56:59 crc kubenswrapper[4980]: I1206 03:56:59.472104 4980 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-nzvgz" Dec 06 03:56:59 crc kubenswrapper[4980]: I1206 03:56:59.537090 4980 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-nzvgz"] Dec 06 03:57:00 crc kubenswrapper[4980]: I1206 03:57:00.341416 4980 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-single-0" Dec 06 03:57:00 crc kubenswrapper[4980]: I1206 03:57:00.376643 4980 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-single-0" Dec 06 03:57:01 crc kubenswrapper[4980]: I1206 03:57:01.477027 4980 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-nzvgz" podUID="d45ee895-b91f-4156-a591-94d6d750a5b2" containerName="registry-server" containerID="cri-o://84ecfc50e7fc781102dbe47be76e32b118fb909f0c51e2274bc301b66e9caab3" gracePeriod=2 Dec 06 03:57:01 crc kubenswrapper[4980]: E1206 03:57:01.576964 4980 secret.go:188] Couldn't get secret glance-kuttl-tests/glance-scripts: secret "glance-scripts" not found Dec 06 03:57:01 crc kubenswrapper[4980]: E1206 03:57:01.577325 4980 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/9f0d1967-5bde-442f-9077-f7871b23df63-scripts podName:9f0d1967-5bde-442f-9077-f7871b23df63 nodeName:}" failed. No retries permitted until 2025-12-06 03:57:02.077297743 +0000 UTC m=+1381.322049024 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "scripts" (UniqueName: "kubernetes.io/secret/9f0d1967-5bde-442f-9077-f7871b23df63-scripts") pod "glance-default-single-0" (UID: "9f0d1967-5bde-442f-9077-f7871b23df63") : secret "glance-scripts" not found Dec 06 03:57:01 crc kubenswrapper[4980]: I1206 03:57:01.582481 4980 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-db-sync-5lft5"] Dec 06 03:57:01 crc kubenswrapper[4980]: I1206 03:57:01.591998 4980 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-db-sync-5lft5"] Dec 06 03:57:01 crc kubenswrapper[4980]: I1206 03:57:01.639714 4980 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance970a-account-delete-mckmw"] Dec 06 03:57:01 crc kubenswrapper[4980]: I1206 03:57:01.641944 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance970a-account-delete-mckmw" Dec 06 03:57:01 crc kubenswrapper[4980]: I1206 03:57:01.665914 4980 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance970a-account-delete-mckmw"] Dec 06 03:57:01 crc kubenswrapper[4980]: I1206 03:57:01.842580 4980 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Dec 06 03:57:01 crc kubenswrapper[4980]: I1206 03:57:01.858653 4980 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-db-create-nxfng"] Dec 06 03:57:01 crc kubenswrapper[4980]: I1206 03:57:01.873585 4980 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-db-create-nxfng"] Dec 06 03:57:01 crc kubenswrapper[4980]: I1206 03:57:01.885709 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g4vzx\" (UniqueName: \"kubernetes.io/projected/7f229f14-48b7-4d6b-a11d-62a9f62598f2-kube-api-access-g4vzx\") pod \"glance970a-account-delete-mckmw\" (UID: \"7f229f14-48b7-4d6b-a11d-62a9f62598f2\") " pod="glance-kuttl-tests/glance970a-account-delete-mckmw" Dec 06 03:57:01 crc kubenswrapper[4980]: I1206 03:57:01.889733 4980 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance970a-account-delete-mckmw"] Dec 06 03:57:01 crc kubenswrapper[4980]: E1206 03:57:01.890313 4980 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[kube-api-access-g4vzx], unattached volumes=[], failed to process volumes=[]: context canceled" pod="glance-kuttl-tests/glance970a-account-delete-mckmw" podUID="7f229f14-48b7-4d6b-a11d-62a9f62598f2" Dec 06 03:57:01 crc kubenswrapper[4980]: I1206 03:57:01.899189 4980 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-970a-account-create-8b64q"] Dec 06 03:57:01 crc kubenswrapper[4980]: I1206 03:57:01.912143 4980 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-970a-account-create-8b64q"] Dec 06 03:57:01 crc kubenswrapper[4980]: I1206 03:57:01.986890 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g4vzx\" (UniqueName: \"kubernetes.io/projected/7f229f14-48b7-4d6b-a11d-62a9f62598f2-kube-api-access-g4vzx\") pod \"glance970a-account-delete-mckmw\" (UID: \"7f229f14-48b7-4d6b-a11d-62a9f62598f2\") " pod="glance-kuttl-tests/glance970a-account-delete-mckmw" Dec 06 03:57:02 crc kubenswrapper[4980]: I1206 03:57:02.005026 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g4vzx\" (UniqueName: \"kubernetes.io/projected/7f229f14-48b7-4d6b-a11d-62a9f62598f2-kube-api-access-g4vzx\") pod \"glance970a-account-delete-mckmw\" (UID: \"7f229f14-48b7-4d6b-a11d-62a9f62598f2\") " pod="glance-kuttl-tests/glance970a-account-delete-mckmw" Dec 06 03:57:02 crc kubenswrapper[4980]: E1206 03:57:02.088125 4980 secret.go:188] Couldn't get secret glance-kuttl-tests/glance-scripts: secret "glance-scripts" not found Dec 06 03:57:02 crc kubenswrapper[4980]: E1206 03:57:02.088207 4980 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/9f0d1967-5bde-442f-9077-f7871b23df63-scripts podName:9f0d1967-5bde-442f-9077-f7871b23df63 nodeName:}" failed. No retries permitted until 2025-12-06 03:57:03.088192929 +0000 UTC m=+1382.332944200 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "scripts" (UniqueName: "kubernetes.io/secret/9f0d1967-5bde-442f-9077-f7871b23df63-scripts") pod "glance-default-single-0" (UID: "9f0d1967-5bde-442f-9077-f7871b23df63") : secret "glance-scripts" not found Dec 06 03:57:02 crc kubenswrapper[4980]: I1206 03:57:02.483466 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance970a-account-delete-mckmw" Dec 06 03:57:02 crc kubenswrapper[4980]: I1206 03:57:02.484941 4980 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-single-0" podUID="9f0d1967-5bde-442f-9077-f7871b23df63" containerName="glance-httpd" containerID="cri-o://03483b8c9a8dd9567feb893a1e8f4e2c8659385420a7a4b0c570edab3f530168" gracePeriod=30 Dec 06 03:57:02 crc kubenswrapper[4980]: I1206 03:57:02.485221 4980 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-single-0" podUID="9f0d1967-5bde-442f-9077-f7871b23df63" containerName="glance-log" containerID="cri-o://84a936e4a197ceb05355a11151840749362c4751bf38242f2c3e953c7ed3e08c" gracePeriod=30 Dec 06 03:57:02 crc kubenswrapper[4980]: I1206 03:57:02.492283 4980 prober.go:107] "Probe failed" probeType="Readiness" pod="glance-kuttl-tests/glance-default-single-0" podUID="9f0d1967-5bde-442f-9077-f7871b23df63" containerName="glance-log" probeResult="failure" output="Get \"https://10.217.0.113:9292/healthcheck\": EOF" Dec 06 03:57:02 crc kubenswrapper[4980]: I1206 03:57:02.501287 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance970a-account-delete-mckmw" Dec 06 03:57:02 crc kubenswrapper[4980]: I1206 03:57:02.697121 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g4vzx\" (UniqueName: \"kubernetes.io/projected/7f229f14-48b7-4d6b-a11d-62a9f62598f2-kube-api-access-g4vzx\") pod \"7f229f14-48b7-4d6b-a11d-62a9f62598f2\" (UID: \"7f229f14-48b7-4d6b-a11d-62a9f62598f2\") " Dec 06 03:57:02 crc kubenswrapper[4980]: I1206 03:57:02.722824 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7f229f14-48b7-4d6b-a11d-62a9f62598f2-kube-api-access-g4vzx" (OuterVolumeSpecName: "kube-api-access-g4vzx") pod "7f229f14-48b7-4d6b-a11d-62a9f62598f2" (UID: "7f229f14-48b7-4d6b-a11d-62a9f62598f2"). InnerVolumeSpecName "kube-api-access-g4vzx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 03:57:02 crc kubenswrapper[4980]: I1206 03:57:02.798616 4980 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g4vzx\" (UniqueName: \"kubernetes.io/projected/7f229f14-48b7-4d6b-a11d-62a9f62598f2-kube-api-access-g4vzx\") on node \"crc\" DevicePath \"\"" Dec 06 03:57:03 crc kubenswrapper[4980]: E1206 03:57:03.106196 4980 secret.go:188] Couldn't get secret glance-kuttl-tests/glance-scripts: secret "glance-scripts" not found Dec 06 03:57:03 crc kubenswrapper[4980]: E1206 03:57:03.106299 4980 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/9f0d1967-5bde-442f-9077-f7871b23df63-scripts podName:9f0d1967-5bde-442f-9077-f7871b23df63 nodeName:}" failed. No retries permitted until 2025-12-06 03:57:05.106271174 +0000 UTC m=+1384.351022485 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "scripts" (UniqueName: "kubernetes.io/secret/9f0d1967-5bde-442f-9077-f7871b23df63-scripts") pod "glance-default-single-0" (UID: "9f0d1967-5bde-442f-9077-f7871b23df63") : secret "glance-scripts" not found Dec 06 03:57:03 crc kubenswrapper[4980]: I1206 03:57:03.193028 4980 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bb09c398-429b-4870-a06b-0ebe969f814e" path="/var/lib/kubelet/pods/bb09c398-429b-4870-a06b-0ebe969f814e/volumes" Dec 06 03:57:03 crc kubenswrapper[4980]: I1206 03:57:03.193794 4980 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d3e30c7b-9b62-40e0-9fca-773b737214a9" path="/var/lib/kubelet/pods/d3e30c7b-9b62-40e0-9fca-773b737214a9/volumes" Dec 06 03:57:03 crc kubenswrapper[4980]: I1206 03:57:03.194302 4980 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fca3de68-d822-49ad-bf87-9b0fab269bbb" path="/var/lib/kubelet/pods/fca3de68-d822-49ad-bf87-9b0fab269bbb/volumes" Dec 06 03:57:03 crc kubenswrapper[4980]: I1206 03:57:03.490458 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance970a-account-delete-mckmw" Dec 06 03:57:03 crc kubenswrapper[4980]: I1206 03:57:03.536938 4980 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance970a-account-delete-mckmw"] Dec 06 03:57:03 crc kubenswrapper[4980]: I1206 03:57:03.546656 4980 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance970a-account-delete-mckmw"] Dec 06 03:57:04 crc kubenswrapper[4980]: I1206 03:57:04.511629 4980 generic.go:334] "Generic (PLEG): container finished" podID="d45ee895-b91f-4156-a591-94d6d750a5b2" containerID="84ecfc50e7fc781102dbe47be76e32b118fb909f0c51e2274bc301b66e9caab3" exitCode=0 Dec 06 03:57:04 crc kubenswrapper[4980]: I1206 03:57:04.511699 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-nzvgz" event={"ID":"d45ee895-b91f-4156-a591-94d6d750a5b2","Type":"ContainerDied","Data":"84ecfc50e7fc781102dbe47be76e32b118fb909f0c51e2274bc301b66e9caab3"} Dec 06 03:57:04 crc kubenswrapper[4980]: I1206 03:57:04.514402 4980 generic.go:334] "Generic (PLEG): container finished" podID="9f0d1967-5bde-442f-9077-f7871b23df63" containerID="84a936e4a197ceb05355a11151840749362c4751bf38242f2c3e953c7ed3e08c" exitCode=143 Dec 06 03:57:04 crc kubenswrapper[4980]: I1206 03:57:04.514452 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"9f0d1967-5bde-442f-9077-f7871b23df63","Type":"ContainerDied","Data":"84a936e4a197ceb05355a11151840749362c4751bf38242f2c3e953c7ed3e08c"} Dec 06 03:57:04 crc kubenswrapper[4980]: I1206 03:57:04.783594 4980 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-nzvgz" Dec 06 03:57:04 crc kubenswrapper[4980]: I1206 03:57:04.965422 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4dtg5\" (UniqueName: \"kubernetes.io/projected/d45ee895-b91f-4156-a591-94d6d750a5b2-kube-api-access-4dtg5\") pod \"d45ee895-b91f-4156-a591-94d6d750a5b2\" (UID: \"d45ee895-b91f-4156-a591-94d6d750a5b2\") " Dec 06 03:57:04 crc kubenswrapper[4980]: I1206 03:57:04.965506 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d45ee895-b91f-4156-a591-94d6d750a5b2-utilities\") pod \"d45ee895-b91f-4156-a591-94d6d750a5b2\" (UID: \"d45ee895-b91f-4156-a591-94d6d750a5b2\") " Dec 06 03:57:04 crc kubenswrapper[4980]: I1206 03:57:04.965553 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d45ee895-b91f-4156-a591-94d6d750a5b2-catalog-content\") pod \"d45ee895-b91f-4156-a591-94d6d750a5b2\" (UID: \"d45ee895-b91f-4156-a591-94d6d750a5b2\") " Dec 06 03:57:04 crc kubenswrapper[4980]: I1206 03:57:04.967232 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d45ee895-b91f-4156-a591-94d6d750a5b2-utilities" (OuterVolumeSpecName: "utilities") pod "d45ee895-b91f-4156-a591-94d6d750a5b2" (UID: "d45ee895-b91f-4156-a591-94d6d750a5b2"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 03:57:04 crc kubenswrapper[4980]: I1206 03:57:04.981794 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d45ee895-b91f-4156-a591-94d6d750a5b2-kube-api-access-4dtg5" (OuterVolumeSpecName: "kube-api-access-4dtg5") pod "d45ee895-b91f-4156-a591-94d6d750a5b2" (UID: "d45ee895-b91f-4156-a591-94d6d750a5b2"). InnerVolumeSpecName "kube-api-access-4dtg5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 03:57:05 crc kubenswrapper[4980]: I1206 03:57:05.067278 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d45ee895-b91f-4156-a591-94d6d750a5b2-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "d45ee895-b91f-4156-a591-94d6d750a5b2" (UID: "d45ee895-b91f-4156-a591-94d6d750a5b2"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 03:57:05 crc kubenswrapper[4980]: I1206 03:57:05.068124 4980 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4dtg5\" (UniqueName: \"kubernetes.io/projected/d45ee895-b91f-4156-a591-94d6d750a5b2-kube-api-access-4dtg5\") on node \"crc\" DevicePath \"\"" Dec 06 03:57:05 crc kubenswrapper[4980]: I1206 03:57:05.068169 4980 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d45ee895-b91f-4156-a591-94d6d750a5b2-utilities\") on node \"crc\" DevicePath \"\"" Dec 06 03:57:05 crc kubenswrapper[4980]: I1206 03:57:05.068186 4980 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d45ee895-b91f-4156-a591-94d6d750a5b2-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 06 03:57:05 crc kubenswrapper[4980]: E1206 03:57:05.169218 4980 secret.go:188] Couldn't get secret glance-kuttl-tests/glance-scripts: secret "glance-scripts" not found Dec 06 03:57:05 crc kubenswrapper[4980]: E1206 03:57:05.169641 4980 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/9f0d1967-5bde-442f-9077-f7871b23df63-scripts podName:9f0d1967-5bde-442f-9077-f7871b23df63 nodeName:}" failed. No retries permitted until 2025-12-06 03:57:09.169612876 +0000 UTC m=+1388.414364167 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "scripts" (UniqueName: "kubernetes.io/secret/9f0d1967-5bde-442f-9077-f7871b23df63-scripts") pod "glance-default-single-0" (UID: "9f0d1967-5bde-442f-9077-f7871b23df63") : secret "glance-scripts" not found Dec 06 03:57:05 crc kubenswrapper[4980]: I1206 03:57:05.194712 4980 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7f229f14-48b7-4d6b-a11d-62a9f62598f2" path="/var/lib/kubelet/pods/7f229f14-48b7-4d6b-a11d-62a9f62598f2/volumes" Dec 06 03:57:05 crc kubenswrapper[4980]: I1206 03:57:05.523567 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-nzvgz" event={"ID":"d45ee895-b91f-4156-a591-94d6d750a5b2","Type":"ContainerDied","Data":"5ecf1a9b020dc94c0e04f3ee99debc6abfe999d6869657d24fb5195b53aa6ba9"} Dec 06 03:57:05 crc kubenswrapper[4980]: I1206 03:57:05.523652 4980 scope.go:117] "RemoveContainer" containerID="84ecfc50e7fc781102dbe47be76e32b118fb909f0c51e2274bc301b66e9caab3" Dec 06 03:57:05 crc kubenswrapper[4980]: I1206 03:57:05.523673 4980 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-nzvgz" Dec 06 03:57:05 crc kubenswrapper[4980]: I1206 03:57:05.549151 4980 scope.go:117] "RemoveContainer" containerID="df7b57e93ae628bf8385c5aec97017b17e1e1b6b4ad00ce9f71415b5afa324eb" Dec 06 03:57:05 crc kubenswrapper[4980]: I1206 03:57:05.549275 4980 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-nzvgz"] Dec 06 03:57:05 crc kubenswrapper[4980]: I1206 03:57:05.555675 4980 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-nzvgz"] Dec 06 03:57:05 crc kubenswrapper[4980]: I1206 03:57:05.570205 4980 scope.go:117] "RemoveContainer" containerID="170f71bef6d6562f39713974265d6189fc745861b19676e72a580d32c9c5e8ce" Dec 06 03:57:06 crc kubenswrapper[4980]: I1206 03:57:06.408468 4980 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-0" Dec 06 03:57:06 crc kubenswrapper[4980]: I1206 03:57:06.512655 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/9f0d1967-5bde-442f-9077-f7871b23df63-public-tls-certs\") pod \"9f0d1967-5bde-442f-9077-f7871b23df63\" (UID: \"9f0d1967-5bde-442f-9077-f7871b23df63\") " Dec 06 03:57:06 crc kubenswrapper[4980]: I1206 03:57:06.512744 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/9f0d1967-5bde-442f-9077-f7871b23df63-internal-tls-certs\") pod \"9f0d1967-5bde-442f-9077-f7871b23df63\" (UID: \"9f0d1967-5bde-442f-9077-f7871b23df63\") " Dec 06 03:57:06 crc kubenswrapper[4980]: I1206 03:57:06.512786 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9f0d1967-5bde-442f-9077-f7871b23df63-config-data\") pod \"9f0d1967-5bde-442f-9077-f7871b23df63\" (UID: \"9f0d1967-5bde-442f-9077-f7871b23df63\") " Dec 06 03:57:06 crc kubenswrapper[4980]: I1206 03:57:06.512831 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9f0d1967-5bde-442f-9077-f7871b23df63-logs\") pod \"9f0d1967-5bde-442f-9077-f7871b23df63\" (UID: \"9f0d1967-5bde-442f-9077-f7871b23df63\") " Dec 06 03:57:06 crc kubenswrapper[4980]: I1206 03:57:06.512885 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9f0d1967-5bde-442f-9077-f7871b23df63-combined-ca-bundle\") pod \"9f0d1967-5bde-442f-9077-f7871b23df63\" (UID: \"9f0d1967-5bde-442f-9077-f7871b23df63\") " Dec 06 03:57:06 crc kubenswrapper[4980]: I1206 03:57:06.513376 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9f0d1967-5bde-442f-9077-f7871b23df63-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "9f0d1967-5bde-442f-9077-f7871b23df63" (UID: "9f0d1967-5bde-442f-9077-f7871b23df63"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 03:57:06 crc kubenswrapper[4980]: I1206 03:57:06.513536 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9f0d1967-5bde-442f-9077-f7871b23df63-logs" (OuterVolumeSpecName: "logs") pod "9f0d1967-5bde-442f-9077-f7871b23df63" (UID: "9f0d1967-5bde-442f-9077-f7871b23df63"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 03:57:06 crc kubenswrapper[4980]: I1206 03:57:06.512914 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/9f0d1967-5bde-442f-9077-f7871b23df63-httpd-run\") pod \"9f0d1967-5bde-442f-9077-f7871b23df63\" (UID: \"9f0d1967-5bde-442f-9077-f7871b23df63\") " Dec 06 03:57:06 crc kubenswrapper[4980]: I1206 03:57:06.513664 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nmsg2\" (UniqueName: \"kubernetes.io/projected/9f0d1967-5bde-442f-9077-f7871b23df63-kube-api-access-nmsg2\") pod \"9f0d1967-5bde-442f-9077-f7871b23df63\" (UID: \"9f0d1967-5bde-442f-9077-f7871b23df63\") " Dec 06 03:57:06 crc kubenswrapper[4980]: I1206 03:57:06.513696 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"9f0d1967-5bde-442f-9077-f7871b23df63\" (UID: \"9f0d1967-5bde-442f-9077-f7871b23df63\") " Dec 06 03:57:06 crc kubenswrapper[4980]: I1206 03:57:06.514209 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9f0d1967-5bde-442f-9077-f7871b23df63-scripts\") pod \"9f0d1967-5bde-442f-9077-f7871b23df63\" (UID: \"9f0d1967-5bde-442f-9077-f7871b23df63\") " Dec 06 03:57:06 crc kubenswrapper[4980]: I1206 03:57:06.514481 4980 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9f0d1967-5bde-442f-9077-f7871b23df63-logs\") on node \"crc\" DevicePath \"\"" Dec 06 03:57:06 crc kubenswrapper[4980]: I1206 03:57:06.514497 4980 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/9f0d1967-5bde-442f-9077-f7871b23df63-httpd-run\") on node \"crc\" DevicePath \"\"" Dec 06 03:57:06 crc kubenswrapper[4980]: I1206 03:57:06.517412 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9f0d1967-5bde-442f-9077-f7871b23df63-kube-api-access-nmsg2" (OuterVolumeSpecName: "kube-api-access-nmsg2") pod "9f0d1967-5bde-442f-9077-f7871b23df63" (UID: "9f0d1967-5bde-442f-9077-f7871b23df63"). InnerVolumeSpecName "kube-api-access-nmsg2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 03:57:06 crc kubenswrapper[4980]: I1206 03:57:06.517814 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage09-crc" (OuterVolumeSpecName: "glance") pod "9f0d1967-5bde-442f-9077-f7871b23df63" (UID: "9f0d1967-5bde-442f-9077-f7871b23df63"). InnerVolumeSpecName "local-storage09-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 06 03:57:06 crc kubenswrapper[4980]: I1206 03:57:06.518204 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9f0d1967-5bde-442f-9077-f7871b23df63-scripts" (OuterVolumeSpecName: "scripts") pod "9f0d1967-5bde-442f-9077-f7871b23df63" (UID: "9f0d1967-5bde-442f-9077-f7871b23df63"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 03:57:06 crc kubenswrapper[4980]: I1206 03:57:06.534108 4980 generic.go:334] "Generic (PLEG): container finished" podID="9f0d1967-5bde-442f-9077-f7871b23df63" containerID="03483b8c9a8dd9567feb893a1e8f4e2c8659385420a7a4b0c570edab3f530168" exitCode=0 Dec 06 03:57:06 crc kubenswrapper[4980]: I1206 03:57:06.534151 4980 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-0" Dec 06 03:57:06 crc kubenswrapper[4980]: I1206 03:57:06.534157 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"9f0d1967-5bde-442f-9077-f7871b23df63","Type":"ContainerDied","Data":"03483b8c9a8dd9567feb893a1e8f4e2c8659385420a7a4b0c570edab3f530168"} Dec 06 03:57:06 crc kubenswrapper[4980]: I1206 03:57:06.534191 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"9f0d1967-5bde-442f-9077-f7871b23df63","Type":"ContainerDied","Data":"0cb082b3657cc6f0e6bc566a3c1bc7de7dcd8314718289e29c88b52bd2525622"} Dec 06 03:57:06 crc kubenswrapper[4980]: I1206 03:57:06.534220 4980 scope.go:117] "RemoveContainer" containerID="03483b8c9a8dd9567feb893a1e8f4e2c8659385420a7a4b0c570edab3f530168" Dec 06 03:57:06 crc kubenswrapper[4980]: I1206 03:57:06.535875 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9f0d1967-5bde-442f-9077-f7871b23df63-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "9f0d1967-5bde-442f-9077-f7871b23df63" (UID: "9f0d1967-5bde-442f-9077-f7871b23df63"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 03:57:06 crc kubenswrapper[4980]: I1206 03:57:06.553245 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9f0d1967-5bde-442f-9077-f7871b23df63-config-data" (OuterVolumeSpecName: "config-data") pod "9f0d1967-5bde-442f-9077-f7871b23df63" (UID: "9f0d1967-5bde-442f-9077-f7871b23df63"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 03:57:06 crc kubenswrapper[4980]: I1206 03:57:06.555444 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9f0d1967-5bde-442f-9077-f7871b23df63-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "9f0d1967-5bde-442f-9077-f7871b23df63" (UID: "9f0d1967-5bde-442f-9077-f7871b23df63"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 03:57:06 crc kubenswrapper[4980]: I1206 03:57:06.562301 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9f0d1967-5bde-442f-9077-f7871b23df63-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "9f0d1967-5bde-442f-9077-f7871b23df63" (UID: "9f0d1967-5bde-442f-9077-f7871b23df63"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 03:57:06 crc kubenswrapper[4980]: I1206 03:57:06.590288 4980 scope.go:117] "RemoveContainer" containerID="84a936e4a197ceb05355a11151840749362c4751bf38242f2c3e953c7ed3e08c" Dec 06 03:57:06 crc kubenswrapper[4980]: I1206 03:57:06.606797 4980 scope.go:117] "RemoveContainer" containerID="03483b8c9a8dd9567feb893a1e8f4e2c8659385420a7a4b0c570edab3f530168" Dec 06 03:57:06 crc kubenswrapper[4980]: E1206 03:57:06.607479 4980 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"03483b8c9a8dd9567feb893a1e8f4e2c8659385420a7a4b0c570edab3f530168\": container with ID starting with 03483b8c9a8dd9567feb893a1e8f4e2c8659385420a7a4b0c570edab3f530168 not found: ID does not exist" containerID="03483b8c9a8dd9567feb893a1e8f4e2c8659385420a7a4b0c570edab3f530168" Dec 06 03:57:06 crc kubenswrapper[4980]: I1206 03:57:06.607573 4980 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"03483b8c9a8dd9567feb893a1e8f4e2c8659385420a7a4b0c570edab3f530168"} err="failed to get container status \"03483b8c9a8dd9567feb893a1e8f4e2c8659385420a7a4b0c570edab3f530168\": rpc error: code = NotFound desc = could not find container \"03483b8c9a8dd9567feb893a1e8f4e2c8659385420a7a4b0c570edab3f530168\": container with ID starting with 03483b8c9a8dd9567feb893a1e8f4e2c8659385420a7a4b0c570edab3f530168 not found: ID does not exist" Dec 06 03:57:06 crc kubenswrapper[4980]: I1206 03:57:06.607622 4980 scope.go:117] "RemoveContainer" containerID="84a936e4a197ceb05355a11151840749362c4751bf38242f2c3e953c7ed3e08c" Dec 06 03:57:06 crc kubenswrapper[4980]: E1206 03:57:06.608138 4980 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"84a936e4a197ceb05355a11151840749362c4751bf38242f2c3e953c7ed3e08c\": container with ID starting with 84a936e4a197ceb05355a11151840749362c4751bf38242f2c3e953c7ed3e08c not found: ID does not exist" containerID="84a936e4a197ceb05355a11151840749362c4751bf38242f2c3e953c7ed3e08c" Dec 06 03:57:06 crc kubenswrapper[4980]: I1206 03:57:06.608236 4980 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"84a936e4a197ceb05355a11151840749362c4751bf38242f2c3e953c7ed3e08c"} err="failed to get container status \"84a936e4a197ceb05355a11151840749362c4751bf38242f2c3e953c7ed3e08c\": rpc error: code = NotFound desc = could not find container \"84a936e4a197ceb05355a11151840749362c4751bf38242f2c3e953c7ed3e08c\": container with ID starting with 84a936e4a197ceb05355a11151840749362c4751bf38242f2c3e953c7ed3e08c not found: ID does not exist" Dec 06 03:57:06 crc kubenswrapper[4980]: I1206 03:57:06.616197 4980 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nmsg2\" (UniqueName: \"kubernetes.io/projected/9f0d1967-5bde-442f-9077-f7871b23df63-kube-api-access-nmsg2\") on node \"crc\" DevicePath \"\"" Dec 06 03:57:06 crc kubenswrapper[4980]: I1206 03:57:06.616319 4980 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") on node \"crc\" " Dec 06 03:57:06 crc kubenswrapper[4980]: I1206 03:57:06.616378 4980 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9f0d1967-5bde-442f-9077-f7871b23df63-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 03:57:06 crc kubenswrapper[4980]: I1206 03:57:06.616441 4980 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/9f0d1967-5bde-442f-9077-f7871b23df63-public-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 06 03:57:06 crc kubenswrapper[4980]: I1206 03:57:06.616497 4980 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/9f0d1967-5bde-442f-9077-f7871b23df63-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 06 03:57:06 crc kubenswrapper[4980]: I1206 03:57:06.616592 4980 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9f0d1967-5bde-442f-9077-f7871b23df63-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 03:57:06 crc kubenswrapper[4980]: I1206 03:57:06.616657 4980 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9f0d1967-5bde-442f-9077-f7871b23df63-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 03:57:06 crc kubenswrapper[4980]: I1206 03:57:06.629658 4980 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage09-crc" (UniqueName: "kubernetes.io/local-volume/local-storage09-crc") on node "crc" Dec 06 03:57:06 crc kubenswrapper[4980]: I1206 03:57:06.718444 4980 reconciler_common.go:293] "Volume detached for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") on node \"crc\" DevicePath \"\"" Dec 06 03:57:06 crc kubenswrapper[4980]: I1206 03:57:06.884852 4980 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Dec 06 03:57:06 crc kubenswrapper[4980]: I1206 03:57:06.891480 4980 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Dec 06 03:57:07 crc kubenswrapper[4980]: I1206 03:57:07.203808 4980 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9f0d1967-5bde-442f-9077-f7871b23df63" path="/var/lib/kubelet/pods/9f0d1967-5bde-442f-9077-f7871b23df63/volumes" Dec 06 03:57:07 crc kubenswrapper[4980]: I1206 03:57:07.204959 4980 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d45ee895-b91f-4156-a591-94d6d750a5b2" path="/var/lib/kubelet/pods/d45ee895-b91f-4156-a591-94d6d750a5b2/volumes" Dec 06 03:57:08 crc kubenswrapper[4980]: I1206 03:57:08.613360 4980 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-db-create-4wjqb"] Dec 06 03:57:08 crc kubenswrapper[4980]: E1206 03:57:08.613847 4980 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d45ee895-b91f-4156-a591-94d6d750a5b2" containerName="extract-utilities" Dec 06 03:57:08 crc kubenswrapper[4980]: I1206 03:57:08.613865 4980 state_mem.go:107] "Deleted CPUSet assignment" podUID="d45ee895-b91f-4156-a591-94d6d750a5b2" containerName="extract-utilities" Dec 06 03:57:08 crc kubenswrapper[4980]: E1206 03:57:08.613878 4980 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d45ee895-b91f-4156-a591-94d6d750a5b2" containerName="extract-content" Dec 06 03:57:08 crc kubenswrapper[4980]: I1206 03:57:08.613886 4980 state_mem.go:107] "Deleted CPUSet assignment" podUID="d45ee895-b91f-4156-a591-94d6d750a5b2" containerName="extract-content" Dec 06 03:57:08 crc kubenswrapper[4980]: E1206 03:57:08.613931 4980 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9f0d1967-5bde-442f-9077-f7871b23df63" containerName="glance-httpd" Dec 06 03:57:08 crc kubenswrapper[4980]: I1206 03:57:08.613941 4980 state_mem.go:107] "Deleted CPUSet assignment" podUID="9f0d1967-5bde-442f-9077-f7871b23df63" containerName="glance-httpd" Dec 06 03:57:08 crc kubenswrapper[4980]: E1206 03:57:08.613960 4980 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d45ee895-b91f-4156-a591-94d6d750a5b2" containerName="registry-server" Dec 06 03:57:08 crc kubenswrapper[4980]: I1206 03:57:08.613968 4980 state_mem.go:107] "Deleted CPUSet assignment" podUID="d45ee895-b91f-4156-a591-94d6d750a5b2" containerName="registry-server" Dec 06 03:57:08 crc kubenswrapper[4980]: E1206 03:57:08.613984 4980 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9f0d1967-5bde-442f-9077-f7871b23df63" containerName="glance-log" Dec 06 03:57:08 crc kubenswrapper[4980]: I1206 03:57:08.613999 4980 state_mem.go:107] "Deleted CPUSet assignment" podUID="9f0d1967-5bde-442f-9077-f7871b23df63" containerName="glance-log" Dec 06 03:57:08 crc kubenswrapper[4980]: I1206 03:57:08.614188 4980 memory_manager.go:354] "RemoveStaleState removing state" podUID="9f0d1967-5bde-442f-9077-f7871b23df63" containerName="glance-log" Dec 06 03:57:08 crc kubenswrapper[4980]: I1206 03:57:08.614215 4980 memory_manager.go:354] "RemoveStaleState removing state" podUID="d45ee895-b91f-4156-a591-94d6d750a5b2" containerName="registry-server" Dec 06 03:57:08 crc kubenswrapper[4980]: I1206 03:57:08.614239 4980 memory_manager.go:354] "RemoveStaleState removing state" podUID="9f0d1967-5bde-442f-9077-f7871b23df63" containerName="glance-httpd" Dec 06 03:57:08 crc kubenswrapper[4980]: I1206 03:57:08.614830 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-create-4wjqb" Dec 06 03:57:08 crc kubenswrapper[4980]: I1206 03:57:08.630443 4980 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-db-create-4wjqb"] Dec 06 03:57:08 crc kubenswrapper[4980]: I1206 03:57:08.752495 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sfq7g\" (UniqueName: \"kubernetes.io/projected/d9f133d0-65f5-4ee3-9906-97e1a883ddf8-kube-api-access-sfq7g\") pod \"glance-db-create-4wjqb\" (UID: \"d9f133d0-65f5-4ee3-9906-97e1a883ddf8\") " pod="glance-kuttl-tests/glance-db-create-4wjqb" Dec 06 03:57:08 crc kubenswrapper[4980]: I1206 03:57:08.854557 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sfq7g\" (UniqueName: \"kubernetes.io/projected/d9f133d0-65f5-4ee3-9906-97e1a883ddf8-kube-api-access-sfq7g\") pod \"glance-db-create-4wjqb\" (UID: \"d9f133d0-65f5-4ee3-9906-97e1a883ddf8\") " pod="glance-kuttl-tests/glance-db-create-4wjqb" Dec 06 03:57:08 crc kubenswrapper[4980]: I1206 03:57:08.887264 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sfq7g\" (UniqueName: \"kubernetes.io/projected/d9f133d0-65f5-4ee3-9906-97e1a883ddf8-kube-api-access-sfq7g\") pod \"glance-db-create-4wjqb\" (UID: \"d9f133d0-65f5-4ee3-9906-97e1a883ddf8\") " pod="glance-kuttl-tests/glance-db-create-4wjqb" Dec 06 03:57:08 crc kubenswrapper[4980]: I1206 03:57:08.947275 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-create-4wjqb" Dec 06 03:57:09 crc kubenswrapper[4980]: I1206 03:57:09.419044 4980 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-db-create-4wjqb"] Dec 06 03:57:09 crc kubenswrapper[4980]: W1206 03:57:09.429985 4980 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd9f133d0_65f5_4ee3_9906_97e1a883ddf8.slice/crio-3b1c54a7630178b51d2d6df42bbad2dcd1ff3fa0184e17859bd94ba83ec2ce38 WatchSource:0}: Error finding container 3b1c54a7630178b51d2d6df42bbad2dcd1ff3fa0184e17859bd94ba83ec2ce38: Status 404 returned error can't find the container with id 3b1c54a7630178b51d2d6df42bbad2dcd1ff3fa0184e17859bd94ba83ec2ce38 Dec 06 03:57:09 crc kubenswrapper[4980]: I1206 03:57:09.564757 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-create-4wjqb" event={"ID":"d9f133d0-65f5-4ee3-9906-97e1a883ddf8","Type":"ContainerStarted","Data":"3b1c54a7630178b51d2d6df42bbad2dcd1ff3fa0184e17859bd94ba83ec2ce38"} Dec 06 03:57:10 crc kubenswrapper[4980]: I1206 03:57:10.577589 4980 generic.go:334] "Generic (PLEG): container finished" podID="d9f133d0-65f5-4ee3-9906-97e1a883ddf8" containerID="4b5687718ac0054924b11fbba5a67c27ce606bd67f575fd23dfa13402b408f11" exitCode=0 Dec 06 03:57:10 crc kubenswrapper[4980]: I1206 03:57:10.577815 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-create-4wjqb" event={"ID":"d9f133d0-65f5-4ee3-9906-97e1a883ddf8","Type":"ContainerDied","Data":"4b5687718ac0054924b11fbba5a67c27ce606bd67f575fd23dfa13402b408f11"} Dec 06 03:57:11 crc kubenswrapper[4980]: I1206 03:57:11.960301 4980 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-create-4wjqb" Dec 06 03:57:12 crc kubenswrapper[4980]: I1206 03:57:12.126627 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sfq7g\" (UniqueName: \"kubernetes.io/projected/d9f133d0-65f5-4ee3-9906-97e1a883ddf8-kube-api-access-sfq7g\") pod \"d9f133d0-65f5-4ee3-9906-97e1a883ddf8\" (UID: \"d9f133d0-65f5-4ee3-9906-97e1a883ddf8\") " Dec 06 03:57:12 crc kubenswrapper[4980]: I1206 03:57:12.139797 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d9f133d0-65f5-4ee3-9906-97e1a883ddf8-kube-api-access-sfq7g" (OuterVolumeSpecName: "kube-api-access-sfq7g") pod "d9f133d0-65f5-4ee3-9906-97e1a883ddf8" (UID: "d9f133d0-65f5-4ee3-9906-97e1a883ddf8"). InnerVolumeSpecName "kube-api-access-sfq7g". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 03:57:12 crc kubenswrapper[4980]: I1206 03:57:12.228773 4980 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sfq7g\" (UniqueName: \"kubernetes.io/projected/d9f133d0-65f5-4ee3-9906-97e1a883ddf8-kube-api-access-sfq7g\") on node \"crc\" DevicePath \"\"" Dec 06 03:57:12 crc kubenswrapper[4980]: I1206 03:57:12.608891 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-create-4wjqb" event={"ID":"d9f133d0-65f5-4ee3-9906-97e1a883ddf8","Type":"ContainerDied","Data":"3b1c54a7630178b51d2d6df42bbad2dcd1ff3fa0184e17859bd94ba83ec2ce38"} Dec 06 03:57:12 crc kubenswrapper[4980]: I1206 03:57:12.608960 4980 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3b1c54a7630178b51d2d6df42bbad2dcd1ff3fa0184e17859bd94ba83ec2ce38" Dec 06 03:57:12 crc kubenswrapper[4980]: I1206 03:57:12.608987 4980 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-create-4wjqb" Dec 06 03:57:18 crc kubenswrapper[4980]: I1206 03:57:18.624725 4980 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-8c70-account-create-mkrvh"] Dec 06 03:57:18 crc kubenswrapper[4980]: E1206 03:57:18.625568 4980 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d9f133d0-65f5-4ee3-9906-97e1a883ddf8" containerName="mariadb-database-create" Dec 06 03:57:18 crc kubenswrapper[4980]: I1206 03:57:18.625582 4980 state_mem.go:107] "Deleted CPUSet assignment" podUID="d9f133d0-65f5-4ee3-9906-97e1a883ddf8" containerName="mariadb-database-create" Dec 06 03:57:18 crc kubenswrapper[4980]: I1206 03:57:18.625733 4980 memory_manager.go:354] "RemoveStaleState removing state" podUID="d9f133d0-65f5-4ee3-9906-97e1a883ddf8" containerName="mariadb-database-create" Dec 06 03:57:18 crc kubenswrapper[4980]: I1206 03:57:18.626173 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-8c70-account-create-mkrvh" Dec 06 03:57:18 crc kubenswrapper[4980]: I1206 03:57:18.638771 4980 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-db-secret" Dec 06 03:57:18 crc kubenswrapper[4980]: I1206 03:57:18.659650 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vpzhv\" (UniqueName: \"kubernetes.io/projected/aae1a12f-6f92-43cc-99ae-fd14a894aadd-kube-api-access-vpzhv\") pod \"glance-8c70-account-create-mkrvh\" (UID: \"aae1a12f-6f92-43cc-99ae-fd14a894aadd\") " pod="glance-kuttl-tests/glance-8c70-account-create-mkrvh" Dec 06 03:57:18 crc kubenswrapper[4980]: I1206 03:57:18.665316 4980 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-8c70-account-create-mkrvh"] Dec 06 03:57:18 crc kubenswrapper[4980]: I1206 03:57:18.760740 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vpzhv\" (UniqueName: \"kubernetes.io/projected/aae1a12f-6f92-43cc-99ae-fd14a894aadd-kube-api-access-vpzhv\") pod \"glance-8c70-account-create-mkrvh\" (UID: \"aae1a12f-6f92-43cc-99ae-fd14a894aadd\") " pod="glance-kuttl-tests/glance-8c70-account-create-mkrvh" Dec 06 03:57:18 crc kubenswrapper[4980]: I1206 03:57:18.779967 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vpzhv\" (UniqueName: \"kubernetes.io/projected/aae1a12f-6f92-43cc-99ae-fd14a894aadd-kube-api-access-vpzhv\") pod \"glance-8c70-account-create-mkrvh\" (UID: \"aae1a12f-6f92-43cc-99ae-fd14a894aadd\") " pod="glance-kuttl-tests/glance-8c70-account-create-mkrvh" Dec 06 03:57:18 crc kubenswrapper[4980]: I1206 03:57:18.943747 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-8c70-account-create-mkrvh" Dec 06 03:57:19 crc kubenswrapper[4980]: I1206 03:57:19.449629 4980 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-8c70-account-create-mkrvh"] Dec 06 03:57:19 crc kubenswrapper[4980]: I1206 03:57:19.676413 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-8c70-account-create-mkrvh" event={"ID":"aae1a12f-6f92-43cc-99ae-fd14a894aadd","Type":"ContainerStarted","Data":"be215c655f9b5a0857ea6486ec2451c3717788171774e8b9f09a21787b8136d9"} Dec 06 03:57:19 crc kubenswrapper[4980]: I1206 03:57:19.676502 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-8c70-account-create-mkrvh" event={"ID":"aae1a12f-6f92-43cc-99ae-fd14a894aadd","Type":"ContainerStarted","Data":"6108b7bbf11d5785064353e67fae99767b072e7318a4138f6cd22f3ec048091a"} Dec 06 03:57:20 crc kubenswrapper[4980]: I1206 03:57:20.689343 4980 generic.go:334] "Generic (PLEG): container finished" podID="aae1a12f-6f92-43cc-99ae-fd14a894aadd" containerID="be215c655f9b5a0857ea6486ec2451c3717788171774e8b9f09a21787b8136d9" exitCode=0 Dec 06 03:57:20 crc kubenswrapper[4980]: I1206 03:57:20.689433 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-8c70-account-create-mkrvh" event={"ID":"aae1a12f-6f92-43cc-99ae-fd14a894aadd","Type":"ContainerDied","Data":"be215c655f9b5a0857ea6486ec2451c3717788171774e8b9f09a21787b8136d9"} Dec 06 03:57:22 crc kubenswrapper[4980]: I1206 03:57:22.050681 4980 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-8c70-account-create-mkrvh" Dec 06 03:57:22 crc kubenswrapper[4980]: I1206 03:57:22.168006 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vpzhv\" (UniqueName: \"kubernetes.io/projected/aae1a12f-6f92-43cc-99ae-fd14a894aadd-kube-api-access-vpzhv\") pod \"aae1a12f-6f92-43cc-99ae-fd14a894aadd\" (UID: \"aae1a12f-6f92-43cc-99ae-fd14a894aadd\") " Dec 06 03:57:22 crc kubenswrapper[4980]: I1206 03:57:22.172988 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/aae1a12f-6f92-43cc-99ae-fd14a894aadd-kube-api-access-vpzhv" (OuterVolumeSpecName: "kube-api-access-vpzhv") pod "aae1a12f-6f92-43cc-99ae-fd14a894aadd" (UID: "aae1a12f-6f92-43cc-99ae-fd14a894aadd"). InnerVolumeSpecName "kube-api-access-vpzhv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 03:57:22 crc kubenswrapper[4980]: I1206 03:57:22.289259 4980 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vpzhv\" (UniqueName: \"kubernetes.io/projected/aae1a12f-6f92-43cc-99ae-fd14a894aadd-kube-api-access-vpzhv\") on node \"crc\" DevicePath \"\"" Dec 06 03:57:22 crc kubenswrapper[4980]: I1206 03:57:22.714648 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-8c70-account-create-mkrvh" event={"ID":"aae1a12f-6f92-43cc-99ae-fd14a894aadd","Type":"ContainerDied","Data":"6108b7bbf11d5785064353e67fae99767b072e7318a4138f6cd22f3ec048091a"} Dec 06 03:57:22 crc kubenswrapper[4980]: I1206 03:57:22.714703 4980 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6108b7bbf11d5785064353e67fae99767b072e7318a4138f6cd22f3ec048091a" Dec 06 03:57:22 crc kubenswrapper[4980]: I1206 03:57:22.714801 4980 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-8c70-account-create-mkrvh" Dec 06 03:57:23 crc kubenswrapper[4980]: I1206 03:57:23.781217 4980 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-db-sync-9lbg6"] Dec 06 03:57:23 crc kubenswrapper[4980]: E1206 03:57:23.781671 4980 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aae1a12f-6f92-43cc-99ae-fd14a894aadd" containerName="mariadb-account-create" Dec 06 03:57:23 crc kubenswrapper[4980]: I1206 03:57:23.781695 4980 state_mem.go:107] "Deleted CPUSet assignment" podUID="aae1a12f-6f92-43cc-99ae-fd14a894aadd" containerName="mariadb-account-create" Dec 06 03:57:23 crc kubenswrapper[4980]: I1206 03:57:23.781945 4980 memory_manager.go:354] "RemoveStaleState removing state" podUID="aae1a12f-6f92-43cc-99ae-fd14a894aadd" containerName="mariadb-account-create" Dec 06 03:57:23 crc kubenswrapper[4980]: I1206 03:57:23.782831 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-sync-9lbg6" Dec 06 03:57:23 crc kubenswrapper[4980]: I1206 03:57:23.785657 4980 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-glance-dockercfg-bmn5t" Dec 06 03:57:23 crc kubenswrapper[4980]: I1206 03:57:23.787287 4980 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-config-data" Dec 06 03:57:23 crc kubenswrapper[4980]: I1206 03:57:23.798953 4980 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-db-sync-9lbg6"] Dec 06 03:57:23 crc kubenswrapper[4980]: I1206 03:57:23.815891 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/34f14e45-81a8-490f-a760-88fbac6f06b3-config-data\") pod \"glance-db-sync-9lbg6\" (UID: \"34f14e45-81a8-490f-a760-88fbac6f06b3\") " pod="glance-kuttl-tests/glance-db-sync-9lbg6" Dec 06 03:57:23 crc kubenswrapper[4980]: I1206 03:57:23.815996 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-25v4f\" (UniqueName: \"kubernetes.io/projected/34f14e45-81a8-490f-a760-88fbac6f06b3-kube-api-access-25v4f\") pod \"glance-db-sync-9lbg6\" (UID: \"34f14e45-81a8-490f-a760-88fbac6f06b3\") " pod="glance-kuttl-tests/glance-db-sync-9lbg6" Dec 06 03:57:23 crc kubenswrapper[4980]: I1206 03:57:23.816161 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/34f14e45-81a8-490f-a760-88fbac6f06b3-db-sync-config-data\") pod \"glance-db-sync-9lbg6\" (UID: \"34f14e45-81a8-490f-a760-88fbac6f06b3\") " pod="glance-kuttl-tests/glance-db-sync-9lbg6" Dec 06 03:57:23 crc kubenswrapper[4980]: I1206 03:57:23.917572 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/34f14e45-81a8-490f-a760-88fbac6f06b3-db-sync-config-data\") pod \"glance-db-sync-9lbg6\" (UID: \"34f14e45-81a8-490f-a760-88fbac6f06b3\") " pod="glance-kuttl-tests/glance-db-sync-9lbg6" Dec 06 03:57:23 crc kubenswrapper[4980]: I1206 03:57:23.917669 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/34f14e45-81a8-490f-a760-88fbac6f06b3-config-data\") pod \"glance-db-sync-9lbg6\" (UID: \"34f14e45-81a8-490f-a760-88fbac6f06b3\") " pod="glance-kuttl-tests/glance-db-sync-9lbg6" Dec 06 03:57:23 crc kubenswrapper[4980]: I1206 03:57:23.917716 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-25v4f\" (UniqueName: \"kubernetes.io/projected/34f14e45-81a8-490f-a760-88fbac6f06b3-kube-api-access-25v4f\") pod \"glance-db-sync-9lbg6\" (UID: \"34f14e45-81a8-490f-a760-88fbac6f06b3\") " pod="glance-kuttl-tests/glance-db-sync-9lbg6" Dec 06 03:57:23 crc kubenswrapper[4980]: I1206 03:57:23.932504 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/34f14e45-81a8-490f-a760-88fbac6f06b3-db-sync-config-data\") pod \"glance-db-sync-9lbg6\" (UID: \"34f14e45-81a8-490f-a760-88fbac6f06b3\") " pod="glance-kuttl-tests/glance-db-sync-9lbg6" Dec 06 03:57:23 crc kubenswrapper[4980]: I1206 03:57:23.944692 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/34f14e45-81a8-490f-a760-88fbac6f06b3-config-data\") pod \"glance-db-sync-9lbg6\" (UID: \"34f14e45-81a8-490f-a760-88fbac6f06b3\") " pod="glance-kuttl-tests/glance-db-sync-9lbg6" Dec 06 03:57:23 crc kubenswrapper[4980]: I1206 03:57:23.950664 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-25v4f\" (UniqueName: \"kubernetes.io/projected/34f14e45-81a8-490f-a760-88fbac6f06b3-kube-api-access-25v4f\") pod \"glance-db-sync-9lbg6\" (UID: \"34f14e45-81a8-490f-a760-88fbac6f06b3\") " pod="glance-kuttl-tests/glance-db-sync-9lbg6" Dec 06 03:57:24 crc kubenswrapper[4980]: I1206 03:57:24.110320 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-sync-9lbg6" Dec 06 03:57:24 crc kubenswrapper[4980]: I1206 03:57:24.533290 4980 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-db-sync-9lbg6"] Dec 06 03:57:24 crc kubenswrapper[4980]: I1206 03:57:24.735755 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-sync-9lbg6" event={"ID":"34f14e45-81a8-490f-a760-88fbac6f06b3","Type":"ContainerStarted","Data":"05923de4b0b3287d59279db041ec8cee7a750fb30c47be76f4f4f4e3f6f810f0"} Dec 06 03:57:25 crc kubenswrapper[4980]: I1206 03:57:25.746586 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-sync-9lbg6" event={"ID":"34f14e45-81a8-490f-a760-88fbac6f06b3","Type":"ContainerStarted","Data":"31ffa96d7d8a317ed02c26ebf3a14d8114b58fb0da24d85117a2c65a3d8b80c7"} Dec 06 03:57:25 crc kubenswrapper[4980]: I1206 03:57:25.784778 4980 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/glance-db-sync-9lbg6" podStartSLOduration=2.784743892 podStartE2EDuration="2.784743892s" podCreationTimestamp="2025-12-06 03:57:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 03:57:25.773479269 +0000 UTC m=+1405.018230540" watchObservedRunningTime="2025-12-06 03:57:25.784743892 +0000 UTC m=+1405.029495163" Dec 06 03:57:27 crc kubenswrapper[4980]: I1206 03:57:27.770601 4980 generic.go:334] "Generic (PLEG): container finished" podID="34f14e45-81a8-490f-a760-88fbac6f06b3" containerID="31ffa96d7d8a317ed02c26ebf3a14d8114b58fb0da24d85117a2c65a3d8b80c7" exitCode=0 Dec 06 03:57:27 crc kubenswrapper[4980]: I1206 03:57:27.770691 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-sync-9lbg6" event={"ID":"34f14e45-81a8-490f-a760-88fbac6f06b3","Type":"ContainerDied","Data":"31ffa96d7d8a317ed02c26ebf3a14d8114b58fb0da24d85117a2c65a3d8b80c7"} Dec 06 03:57:29 crc kubenswrapper[4980]: I1206 03:57:29.105794 4980 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-sync-9lbg6" Dec 06 03:57:29 crc kubenswrapper[4980]: I1206 03:57:29.302025 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-25v4f\" (UniqueName: \"kubernetes.io/projected/34f14e45-81a8-490f-a760-88fbac6f06b3-kube-api-access-25v4f\") pod \"34f14e45-81a8-490f-a760-88fbac6f06b3\" (UID: \"34f14e45-81a8-490f-a760-88fbac6f06b3\") " Dec 06 03:57:29 crc kubenswrapper[4980]: I1206 03:57:29.302104 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/34f14e45-81a8-490f-a760-88fbac6f06b3-db-sync-config-data\") pod \"34f14e45-81a8-490f-a760-88fbac6f06b3\" (UID: \"34f14e45-81a8-490f-a760-88fbac6f06b3\") " Dec 06 03:57:29 crc kubenswrapper[4980]: I1206 03:57:29.302251 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/34f14e45-81a8-490f-a760-88fbac6f06b3-config-data\") pod \"34f14e45-81a8-490f-a760-88fbac6f06b3\" (UID: \"34f14e45-81a8-490f-a760-88fbac6f06b3\") " Dec 06 03:57:29 crc kubenswrapper[4980]: I1206 03:57:29.313224 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/34f14e45-81a8-490f-a760-88fbac6f06b3-kube-api-access-25v4f" (OuterVolumeSpecName: "kube-api-access-25v4f") pod "34f14e45-81a8-490f-a760-88fbac6f06b3" (UID: "34f14e45-81a8-490f-a760-88fbac6f06b3"). InnerVolumeSpecName "kube-api-access-25v4f". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 03:57:29 crc kubenswrapper[4980]: I1206 03:57:29.314283 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/34f14e45-81a8-490f-a760-88fbac6f06b3-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "34f14e45-81a8-490f-a760-88fbac6f06b3" (UID: "34f14e45-81a8-490f-a760-88fbac6f06b3"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 03:57:29 crc kubenswrapper[4980]: I1206 03:57:29.356317 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/34f14e45-81a8-490f-a760-88fbac6f06b3-config-data" (OuterVolumeSpecName: "config-data") pod "34f14e45-81a8-490f-a760-88fbac6f06b3" (UID: "34f14e45-81a8-490f-a760-88fbac6f06b3"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 03:57:29 crc kubenswrapper[4980]: I1206 03:57:29.404455 4980 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-25v4f\" (UniqueName: \"kubernetes.io/projected/34f14e45-81a8-490f-a760-88fbac6f06b3-kube-api-access-25v4f\") on node \"crc\" DevicePath \"\"" Dec 06 03:57:29 crc kubenswrapper[4980]: I1206 03:57:29.404885 4980 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/34f14e45-81a8-490f-a760-88fbac6f06b3-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 03:57:29 crc kubenswrapper[4980]: I1206 03:57:29.404913 4980 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/34f14e45-81a8-490f-a760-88fbac6f06b3-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 03:57:29 crc kubenswrapper[4980]: I1206 03:57:29.797270 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-sync-9lbg6" event={"ID":"34f14e45-81a8-490f-a760-88fbac6f06b3","Type":"ContainerDied","Data":"05923de4b0b3287d59279db041ec8cee7a750fb30c47be76f4f4f4e3f6f810f0"} Dec 06 03:57:29 crc kubenswrapper[4980]: I1206 03:57:29.797328 4980 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="05923de4b0b3287d59279db041ec8cee7a750fb30c47be76f4f4f4e3f6f810f0" Dec 06 03:57:29 crc kubenswrapper[4980]: I1206 03:57:29.797335 4980 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-sync-9lbg6" Dec 06 03:57:31 crc kubenswrapper[4980]: I1206 03:57:31.169768 4980 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-default-external-api-0"] Dec 06 03:57:31 crc kubenswrapper[4980]: E1206 03:57:31.170130 4980 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="34f14e45-81a8-490f-a760-88fbac6f06b3" containerName="glance-db-sync" Dec 06 03:57:31 crc kubenswrapper[4980]: I1206 03:57:31.170148 4980 state_mem.go:107] "Deleted CPUSet assignment" podUID="34f14e45-81a8-490f-a760-88fbac6f06b3" containerName="glance-db-sync" Dec 06 03:57:31 crc kubenswrapper[4980]: I1206 03:57:31.170288 4980 memory_manager.go:354] "RemoveStaleState removing state" podUID="34f14e45-81a8-490f-a760-88fbac6f06b3" containerName="glance-db-sync" Dec 06 03:57:31 crc kubenswrapper[4980]: I1206 03:57:31.171473 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-external-api-0" Dec 06 03:57:31 crc kubenswrapper[4980]: I1206 03:57:31.173595 4980 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-scripts" Dec 06 03:57:31 crc kubenswrapper[4980]: I1206 03:57:31.173624 4980 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-glance-dockercfg-bmn5t" Dec 06 03:57:31 crc kubenswrapper[4980]: I1206 03:57:31.174150 4980 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-default-external-config-data" Dec 06 03:57:31 crc kubenswrapper[4980]: I1206 03:57:31.193111 4980 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-external-api-0"] Dec 06 03:57:31 crc kubenswrapper[4980]: I1206 03:57:31.337106 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/7d66b977-3b53-4fe4-bc08-9e9083b32d76-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"7d66b977-3b53-4fe4-bc08-9e9083b32d76\") " pod="glance-kuttl-tests/glance-default-external-api-0" Dec 06 03:57:31 crc kubenswrapper[4980]: I1206 03:57:31.337177 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/7d66b977-3b53-4fe4-bc08-9e9083b32d76-run\") pod \"glance-default-external-api-0\" (UID: \"7d66b977-3b53-4fe4-bc08-9e9083b32d76\") " pod="glance-kuttl-tests/glance-default-external-api-0" Dec 06 03:57:31 crc kubenswrapper[4980]: I1206 03:57:31.337230 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/7d66b977-3b53-4fe4-bc08-9e9083b32d76-etc-iscsi\") pod \"glance-default-external-api-0\" (UID: \"7d66b977-3b53-4fe4-bc08-9e9083b32d76\") " pod="glance-kuttl-tests/glance-default-external-api-0" Dec 06 03:57:31 crc kubenswrapper[4980]: I1206 03:57:31.337257 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/7d66b977-3b53-4fe4-bc08-9e9083b32d76-sys\") pod \"glance-default-external-api-0\" (UID: \"7d66b977-3b53-4fe4-bc08-9e9083b32d76\") " pod="glance-kuttl-tests/glance-default-external-api-0" Dec 06 03:57:31 crc kubenswrapper[4980]: I1206 03:57:31.337291 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-external-api-0\" (UID: \"7d66b977-3b53-4fe4-bc08-9e9083b32d76\") " pod="glance-kuttl-tests/glance-default-external-api-0" Dec 06 03:57:31 crc kubenswrapper[4980]: I1206 03:57:31.337321 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/7d66b977-3b53-4fe4-bc08-9e9083b32d76-lib-modules\") pod \"glance-default-external-api-0\" (UID: \"7d66b977-3b53-4fe4-bc08-9e9083b32d76\") " pod="glance-kuttl-tests/glance-default-external-api-0" Dec 06 03:57:31 crc kubenswrapper[4980]: I1206 03:57:31.337351 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7d66b977-3b53-4fe4-bc08-9e9083b32d76-scripts\") pod \"glance-default-external-api-0\" (UID: \"7d66b977-3b53-4fe4-bc08-9e9083b32d76\") " pod="glance-kuttl-tests/glance-default-external-api-0" Dec 06 03:57:31 crc kubenswrapper[4980]: I1206 03:57:31.337379 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/7d66b977-3b53-4fe4-bc08-9e9083b32d76-etc-nvme\") pod \"glance-default-external-api-0\" (UID: \"7d66b977-3b53-4fe4-bc08-9e9083b32d76\") " pod="glance-kuttl-tests/glance-default-external-api-0" Dec 06 03:57:31 crc kubenswrapper[4980]: I1206 03:57:31.337438 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lrtc2\" (UniqueName: \"kubernetes.io/projected/7d66b977-3b53-4fe4-bc08-9e9083b32d76-kube-api-access-lrtc2\") pod \"glance-default-external-api-0\" (UID: \"7d66b977-3b53-4fe4-bc08-9e9083b32d76\") " pod="glance-kuttl-tests/glance-default-external-api-0" Dec 06 03:57:31 crc kubenswrapper[4980]: I1206 03:57:31.337480 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7d66b977-3b53-4fe4-bc08-9e9083b32d76-config-data\") pod \"glance-default-external-api-0\" (UID: \"7d66b977-3b53-4fe4-bc08-9e9083b32d76\") " pod="glance-kuttl-tests/glance-default-external-api-0" Dec 06 03:57:31 crc kubenswrapper[4980]: I1206 03:57:31.337590 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/7d66b977-3b53-4fe4-bc08-9e9083b32d76-var-locks-brick\") pod \"glance-default-external-api-0\" (UID: \"7d66b977-3b53-4fe4-bc08-9e9083b32d76\") " pod="glance-kuttl-tests/glance-default-external-api-0" Dec 06 03:57:31 crc kubenswrapper[4980]: I1206 03:57:31.338299 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/7d66b977-3b53-4fe4-bc08-9e9083b32d76-dev\") pod \"glance-default-external-api-0\" (UID: \"7d66b977-3b53-4fe4-bc08-9e9083b32d76\") " pod="glance-kuttl-tests/glance-default-external-api-0" Dec 06 03:57:31 crc kubenswrapper[4980]: I1206 03:57:31.338346 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7d66b977-3b53-4fe4-bc08-9e9083b32d76-logs\") pod \"glance-default-external-api-0\" (UID: \"7d66b977-3b53-4fe4-bc08-9e9083b32d76\") " pod="glance-kuttl-tests/glance-default-external-api-0" Dec 06 03:57:31 crc kubenswrapper[4980]: I1206 03:57:31.338367 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-external-api-0\" (UID: \"7d66b977-3b53-4fe4-bc08-9e9083b32d76\") " pod="glance-kuttl-tests/glance-default-external-api-0" Dec 06 03:57:31 crc kubenswrapper[4980]: I1206 03:57:31.361278 4980 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-0"] Dec 06 03:57:31 crc kubenswrapper[4980]: I1206 03:57:31.362668 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 06 03:57:31 crc kubenswrapper[4980]: I1206 03:57:31.365075 4980 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-default-internal-config-data" Dec 06 03:57:31 crc kubenswrapper[4980]: I1206 03:57:31.377938 4980 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-0"] Dec 06 03:57:31 crc kubenswrapper[4980]: I1206 03:57:31.440242 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-external-api-0\" (UID: \"7d66b977-3b53-4fe4-bc08-9e9083b32d76\") " pod="glance-kuttl-tests/glance-default-external-api-0" Dec 06 03:57:31 crc kubenswrapper[4980]: I1206 03:57:31.440594 4980 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-external-api-0\" (UID: \"7d66b977-3b53-4fe4-bc08-9e9083b32d76\") device mount path \"/mnt/openstack/pv10\"" pod="glance-kuttl-tests/glance-default-external-api-0" Dec 06 03:57:31 crc kubenswrapper[4980]: I1206 03:57:31.440606 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/7d66b977-3b53-4fe4-bc08-9e9083b32d76-lib-modules\") pod \"glance-default-external-api-0\" (UID: \"7d66b977-3b53-4fe4-bc08-9e9083b32d76\") " pod="glance-kuttl-tests/glance-default-external-api-0" Dec 06 03:57:31 crc kubenswrapper[4980]: I1206 03:57:31.440637 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7d66b977-3b53-4fe4-bc08-9e9083b32d76-scripts\") pod \"glance-default-external-api-0\" (UID: \"7d66b977-3b53-4fe4-bc08-9e9083b32d76\") " pod="glance-kuttl-tests/glance-default-external-api-0" Dec 06 03:57:31 crc kubenswrapper[4980]: I1206 03:57:31.440661 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/7d66b977-3b53-4fe4-bc08-9e9083b32d76-etc-nvme\") pod \"glance-default-external-api-0\" (UID: \"7d66b977-3b53-4fe4-bc08-9e9083b32d76\") " pod="glance-kuttl-tests/glance-default-external-api-0" Dec 06 03:57:31 crc kubenswrapper[4980]: I1206 03:57:31.440691 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lrtc2\" (UniqueName: \"kubernetes.io/projected/7d66b977-3b53-4fe4-bc08-9e9083b32d76-kube-api-access-lrtc2\") pod \"glance-default-external-api-0\" (UID: \"7d66b977-3b53-4fe4-bc08-9e9083b32d76\") " pod="glance-kuttl-tests/glance-default-external-api-0" Dec 06 03:57:31 crc kubenswrapper[4980]: I1206 03:57:31.440702 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/7d66b977-3b53-4fe4-bc08-9e9083b32d76-lib-modules\") pod \"glance-default-external-api-0\" (UID: \"7d66b977-3b53-4fe4-bc08-9e9083b32d76\") " pod="glance-kuttl-tests/glance-default-external-api-0" Dec 06 03:57:31 crc kubenswrapper[4980]: I1206 03:57:31.440723 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7d66b977-3b53-4fe4-bc08-9e9083b32d76-config-data\") pod \"glance-default-external-api-0\" (UID: \"7d66b977-3b53-4fe4-bc08-9e9083b32d76\") " pod="glance-kuttl-tests/glance-default-external-api-0" Dec 06 03:57:31 crc kubenswrapper[4980]: I1206 03:57:31.440925 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/7d66b977-3b53-4fe4-bc08-9e9083b32d76-etc-nvme\") pod \"glance-default-external-api-0\" (UID: \"7d66b977-3b53-4fe4-bc08-9e9083b32d76\") " pod="glance-kuttl-tests/glance-default-external-api-0" Dec 06 03:57:31 crc kubenswrapper[4980]: I1206 03:57:31.440935 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/7d66b977-3b53-4fe4-bc08-9e9083b32d76-var-locks-brick\") pod \"glance-default-external-api-0\" (UID: \"7d66b977-3b53-4fe4-bc08-9e9083b32d76\") " pod="glance-kuttl-tests/glance-default-external-api-0" Dec 06 03:57:31 crc kubenswrapper[4980]: I1206 03:57:31.440998 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/7d66b977-3b53-4fe4-bc08-9e9083b32d76-dev\") pod \"glance-default-external-api-0\" (UID: \"7d66b977-3b53-4fe4-bc08-9e9083b32d76\") " pod="glance-kuttl-tests/glance-default-external-api-0" Dec 06 03:57:31 crc kubenswrapper[4980]: I1206 03:57:31.441006 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/7d66b977-3b53-4fe4-bc08-9e9083b32d76-var-locks-brick\") pod \"glance-default-external-api-0\" (UID: \"7d66b977-3b53-4fe4-bc08-9e9083b32d76\") " pod="glance-kuttl-tests/glance-default-external-api-0" Dec 06 03:57:31 crc kubenswrapper[4980]: I1206 03:57:31.441025 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7d66b977-3b53-4fe4-bc08-9e9083b32d76-logs\") pod \"glance-default-external-api-0\" (UID: \"7d66b977-3b53-4fe4-bc08-9e9083b32d76\") " pod="glance-kuttl-tests/glance-default-external-api-0" Dec 06 03:57:31 crc kubenswrapper[4980]: I1206 03:57:31.441034 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/7d66b977-3b53-4fe4-bc08-9e9083b32d76-dev\") pod \"glance-default-external-api-0\" (UID: \"7d66b977-3b53-4fe4-bc08-9e9083b32d76\") " pod="glance-kuttl-tests/glance-default-external-api-0" Dec 06 03:57:31 crc kubenswrapper[4980]: I1206 03:57:31.441050 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-external-api-0\" (UID: \"7d66b977-3b53-4fe4-bc08-9e9083b32d76\") " pod="glance-kuttl-tests/glance-default-external-api-0" Dec 06 03:57:31 crc kubenswrapper[4980]: I1206 03:57:31.441088 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/7d66b977-3b53-4fe4-bc08-9e9083b32d76-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"7d66b977-3b53-4fe4-bc08-9e9083b32d76\") " pod="glance-kuttl-tests/glance-default-external-api-0" Dec 06 03:57:31 crc kubenswrapper[4980]: I1206 03:57:31.441115 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/7d66b977-3b53-4fe4-bc08-9e9083b32d76-run\") pod \"glance-default-external-api-0\" (UID: \"7d66b977-3b53-4fe4-bc08-9e9083b32d76\") " pod="glance-kuttl-tests/glance-default-external-api-0" Dec 06 03:57:31 crc kubenswrapper[4980]: I1206 03:57:31.441151 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/7d66b977-3b53-4fe4-bc08-9e9083b32d76-etc-iscsi\") pod \"glance-default-external-api-0\" (UID: \"7d66b977-3b53-4fe4-bc08-9e9083b32d76\") " pod="glance-kuttl-tests/glance-default-external-api-0" Dec 06 03:57:31 crc kubenswrapper[4980]: I1206 03:57:31.441169 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/7d66b977-3b53-4fe4-bc08-9e9083b32d76-sys\") pod \"glance-default-external-api-0\" (UID: \"7d66b977-3b53-4fe4-bc08-9e9083b32d76\") " pod="glance-kuttl-tests/glance-default-external-api-0" Dec 06 03:57:31 crc kubenswrapper[4980]: I1206 03:57:31.441232 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/7d66b977-3b53-4fe4-bc08-9e9083b32d76-sys\") pod \"glance-default-external-api-0\" (UID: \"7d66b977-3b53-4fe4-bc08-9e9083b32d76\") " pod="glance-kuttl-tests/glance-default-external-api-0" Dec 06 03:57:31 crc kubenswrapper[4980]: I1206 03:57:31.441490 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/7d66b977-3b53-4fe4-bc08-9e9083b32d76-run\") pod \"glance-default-external-api-0\" (UID: \"7d66b977-3b53-4fe4-bc08-9e9083b32d76\") " pod="glance-kuttl-tests/glance-default-external-api-0" Dec 06 03:57:31 crc kubenswrapper[4980]: I1206 03:57:31.441530 4980 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-external-api-0\" (UID: \"7d66b977-3b53-4fe4-bc08-9e9083b32d76\") device mount path \"/mnt/openstack/pv11\"" pod="glance-kuttl-tests/glance-default-external-api-0" Dec 06 03:57:31 crc kubenswrapper[4980]: I1206 03:57:31.441568 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/7d66b977-3b53-4fe4-bc08-9e9083b32d76-etc-iscsi\") pod \"glance-default-external-api-0\" (UID: \"7d66b977-3b53-4fe4-bc08-9e9083b32d76\") " pod="glance-kuttl-tests/glance-default-external-api-0" Dec 06 03:57:31 crc kubenswrapper[4980]: I1206 03:57:31.441726 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7d66b977-3b53-4fe4-bc08-9e9083b32d76-logs\") pod \"glance-default-external-api-0\" (UID: \"7d66b977-3b53-4fe4-bc08-9e9083b32d76\") " pod="glance-kuttl-tests/glance-default-external-api-0" Dec 06 03:57:31 crc kubenswrapper[4980]: I1206 03:57:31.441749 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/7d66b977-3b53-4fe4-bc08-9e9083b32d76-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"7d66b977-3b53-4fe4-bc08-9e9083b32d76\") " pod="glance-kuttl-tests/glance-default-external-api-0" Dec 06 03:57:31 crc kubenswrapper[4980]: I1206 03:57:31.449100 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7d66b977-3b53-4fe4-bc08-9e9083b32d76-config-data\") pod \"glance-default-external-api-0\" (UID: \"7d66b977-3b53-4fe4-bc08-9e9083b32d76\") " pod="glance-kuttl-tests/glance-default-external-api-0" Dec 06 03:57:31 crc kubenswrapper[4980]: I1206 03:57:31.449878 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7d66b977-3b53-4fe4-bc08-9e9083b32d76-scripts\") pod \"glance-default-external-api-0\" (UID: \"7d66b977-3b53-4fe4-bc08-9e9083b32d76\") " pod="glance-kuttl-tests/glance-default-external-api-0" Dec 06 03:57:31 crc kubenswrapper[4980]: I1206 03:57:31.471778 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-external-api-0\" (UID: \"7d66b977-3b53-4fe4-bc08-9e9083b32d76\") " pod="glance-kuttl-tests/glance-default-external-api-0" Dec 06 03:57:31 crc kubenswrapper[4980]: I1206 03:57:31.478308 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-external-api-0\" (UID: \"7d66b977-3b53-4fe4-bc08-9e9083b32d76\") " pod="glance-kuttl-tests/glance-default-external-api-0" Dec 06 03:57:31 crc kubenswrapper[4980]: I1206 03:57:31.481007 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lrtc2\" (UniqueName: \"kubernetes.io/projected/7d66b977-3b53-4fe4-bc08-9e9083b32d76-kube-api-access-lrtc2\") pod \"glance-default-external-api-0\" (UID: \"7d66b977-3b53-4fe4-bc08-9e9083b32d76\") " pod="glance-kuttl-tests/glance-default-external-api-0" Dec 06 03:57:31 crc kubenswrapper[4980]: I1206 03:57:31.486745 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-external-api-0" Dec 06 03:57:31 crc kubenswrapper[4980]: I1206 03:57:31.542631 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f5jx4\" (UniqueName: \"kubernetes.io/projected/7617925e-d725-4eec-8ea3-7d9eed28d8d7-kube-api-access-f5jx4\") pod \"glance-default-internal-api-0\" (UID: \"7617925e-d725-4eec-8ea3-7d9eed28d8d7\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 06 03:57:31 crc kubenswrapper[4980]: I1206 03:57:31.542703 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/7617925e-d725-4eec-8ea3-7d9eed28d8d7-etc-nvme\") pod \"glance-default-internal-api-0\" (UID: \"7617925e-d725-4eec-8ea3-7d9eed28d8d7\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 06 03:57:31 crc kubenswrapper[4980]: I1206 03:57:31.542735 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7617925e-d725-4eec-8ea3-7d9eed28d8d7-config-data\") pod \"glance-default-internal-api-0\" (UID: \"7617925e-d725-4eec-8ea3-7d9eed28d8d7\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 06 03:57:31 crc kubenswrapper[4980]: I1206 03:57:31.542759 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/7617925e-d725-4eec-8ea3-7d9eed28d8d7-run\") pod \"glance-default-internal-api-0\" (UID: \"7617925e-d725-4eec-8ea3-7d9eed28d8d7\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 06 03:57:31 crc kubenswrapper[4980]: I1206 03:57:31.542786 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-internal-api-0\" (UID: \"7617925e-d725-4eec-8ea3-7d9eed28d8d7\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 06 03:57:31 crc kubenswrapper[4980]: I1206 03:57:31.542807 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/7617925e-d725-4eec-8ea3-7d9eed28d8d7-etc-iscsi\") pod \"glance-default-internal-api-0\" (UID: \"7617925e-d725-4eec-8ea3-7d9eed28d8d7\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 06 03:57:31 crc kubenswrapper[4980]: I1206 03:57:31.542829 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/7617925e-d725-4eec-8ea3-7d9eed28d8d7-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"7617925e-d725-4eec-8ea3-7d9eed28d8d7\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 06 03:57:31 crc kubenswrapper[4980]: I1206 03:57:31.542870 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/7617925e-d725-4eec-8ea3-7d9eed28d8d7-dev\") pod \"glance-default-internal-api-0\" (UID: \"7617925e-d725-4eec-8ea3-7d9eed28d8d7\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 06 03:57:31 crc kubenswrapper[4980]: I1206 03:57:31.542893 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-internal-api-0\" (UID: \"7617925e-d725-4eec-8ea3-7d9eed28d8d7\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 06 03:57:31 crc kubenswrapper[4980]: I1206 03:57:31.542931 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7617925e-d725-4eec-8ea3-7d9eed28d8d7-logs\") pod \"glance-default-internal-api-0\" (UID: \"7617925e-d725-4eec-8ea3-7d9eed28d8d7\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 06 03:57:31 crc kubenswrapper[4980]: I1206 03:57:31.542957 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7617925e-d725-4eec-8ea3-7d9eed28d8d7-scripts\") pod \"glance-default-internal-api-0\" (UID: \"7617925e-d725-4eec-8ea3-7d9eed28d8d7\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 06 03:57:31 crc kubenswrapper[4980]: I1206 03:57:31.542983 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/7617925e-d725-4eec-8ea3-7d9eed28d8d7-lib-modules\") pod \"glance-default-internal-api-0\" (UID: \"7617925e-d725-4eec-8ea3-7d9eed28d8d7\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 06 03:57:31 crc kubenswrapper[4980]: I1206 03:57:31.543003 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/7617925e-d725-4eec-8ea3-7d9eed28d8d7-sys\") pod \"glance-default-internal-api-0\" (UID: \"7617925e-d725-4eec-8ea3-7d9eed28d8d7\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 06 03:57:31 crc kubenswrapper[4980]: I1206 03:57:31.543028 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/7617925e-d725-4eec-8ea3-7d9eed28d8d7-var-locks-brick\") pod \"glance-default-internal-api-0\" (UID: \"7617925e-d725-4eec-8ea3-7d9eed28d8d7\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 06 03:57:31 crc kubenswrapper[4980]: I1206 03:57:31.644745 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7617925e-d725-4eec-8ea3-7d9eed28d8d7-config-data\") pod \"glance-default-internal-api-0\" (UID: \"7617925e-d725-4eec-8ea3-7d9eed28d8d7\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 06 03:57:31 crc kubenswrapper[4980]: I1206 03:57:31.645015 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/7617925e-d725-4eec-8ea3-7d9eed28d8d7-run\") pod \"glance-default-internal-api-0\" (UID: \"7617925e-d725-4eec-8ea3-7d9eed28d8d7\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 06 03:57:31 crc kubenswrapper[4980]: I1206 03:57:31.645043 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-internal-api-0\" (UID: \"7617925e-d725-4eec-8ea3-7d9eed28d8d7\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 06 03:57:31 crc kubenswrapper[4980]: I1206 03:57:31.645057 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/7617925e-d725-4eec-8ea3-7d9eed28d8d7-etc-iscsi\") pod \"glance-default-internal-api-0\" (UID: \"7617925e-d725-4eec-8ea3-7d9eed28d8d7\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 06 03:57:31 crc kubenswrapper[4980]: I1206 03:57:31.645075 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/7617925e-d725-4eec-8ea3-7d9eed28d8d7-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"7617925e-d725-4eec-8ea3-7d9eed28d8d7\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 06 03:57:31 crc kubenswrapper[4980]: I1206 03:57:31.645109 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/7617925e-d725-4eec-8ea3-7d9eed28d8d7-dev\") pod \"glance-default-internal-api-0\" (UID: \"7617925e-d725-4eec-8ea3-7d9eed28d8d7\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 06 03:57:31 crc kubenswrapper[4980]: I1206 03:57:31.645125 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-internal-api-0\" (UID: \"7617925e-d725-4eec-8ea3-7d9eed28d8d7\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 06 03:57:31 crc kubenswrapper[4980]: I1206 03:57:31.645154 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7617925e-d725-4eec-8ea3-7d9eed28d8d7-logs\") pod \"glance-default-internal-api-0\" (UID: \"7617925e-d725-4eec-8ea3-7d9eed28d8d7\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 06 03:57:31 crc kubenswrapper[4980]: I1206 03:57:31.645172 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7617925e-d725-4eec-8ea3-7d9eed28d8d7-scripts\") pod \"glance-default-internal-api-0\" (UID: \"7617925e-d725-4eec-8ea3-7d9eed28d8d7\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 06 03:57:31 crc kubenswrapper[4980]: I1206 03:57:31.645189 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/7617925e-d725-4eec-8ea3-7d9eed28d8d7-lib-modules\") pod \"glance-default-internal-api-0\" (UID: \"7617925e-d725-4eec-8ea3-7d9eed28d8d7\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 06 03:57:31 crc kubenswrapper[4980]: I1206 03:57:31.645205 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/7617925e-d725-4eec-8ea3-7d9eed28d8d7-sys\") pod \"glance-default-internal-api-0\" (UID: \"7617925e-d725-4eec-8ea3-7d9eed28d8d7\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 06 03:57:31 crc kubenswrapper[4980]: I1206 03:57:31.645223 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/7617925e-d725-4eec-8ea3-7d9eed28d8d7-var-locks-brick\") pod \"glance-default-internal-api-0\" (UID: \"7617925e-d725-4eec-8ea3-7d9eed28d8d7\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 06 03:57:31 crc kubenswrapper[4980]: I1206 03:57:31.645268 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f5jx4\" (UniqueName: \"kubernetes.io/projected/7617925e-d725-4eec-8ea3-7d9eed28d8d7-kube-api-access-f5jx4\") pod \"glance-default-internal-api-0\" (UID: \"7617925e-d725-4eec-8ea3-7d9eed28d8d7\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 06 03:57:31 crc kubenswrapper[4980]: I1206 03:57:31.645290 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/7617925e-d725-4eec-8ea3-7d9eed28d8d7-etc-nvme\") pod \"glance-default-internal-api-0\" (UID: \"7617925e-d725-4eec-8ea3-7d9eed28d8d7\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 06 03:57:31 crc kubenswrapper[4980]: I1206 03:57:31.645309 4980 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-internal-api-0\" (UID: \"7617925e-d725-4eec-8ea3-7d9eed28d8d7\") device mount path \"/mnt/openstack/pv12\"" pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 06 03:57:31 crc kubenswrapper[4980]: I1206 03:57:31.645433 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/7617925e-d725-4eec-8ea3-7d9eed28d8d7-etc-nvme\") pod \"glance-default-internal-api-0\" (UID: \"7617925e-d725-4eec-8ea3-7d9eed28d8d7\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 06 03:57:31 crc kubenswrapper[4980]: I1206 03:57:31.645470 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/7617925e-d725-4eec-8ea3-7d9eed28d8d7-etc-iscsi\") pod \"glance-default-internal-api-0\" (UID: \"7617925e-d725-4eec-8ea3-7d9eed28d8d7\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 06 03:57:31 crc kubenswrapper[4980]: I1206 03:57:31.645833 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/7617925e-d725-4eec-8ea3-7d9eed28d8d7-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"7617925e-d725-4eec-8ea3-7d9eed28d8d7\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 06 03:57:31 crc kubenswrapper[4980]: I1206 03:57:31.645904 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/7617925e-d725-4eec-8ea3-7d9eed28d8d7-dev\") pod \"glance-default-internal-api-0\" (UID: \"7617925e-d725-4eec-8ea3-7d9eed28d8d7\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 06 03:57:31 crc kubenswrapper[4980]: I1206 03:57:31.645992 4980 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-internal-api-0\" (UID: \"7617925e-d725-4eec-8ea3-7d9eed28d8d7\") device mount path \"/mnt/openstack/pv01\"" pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 06 03:57:31 crc kubenswrapper[4980]: I1206 03:57:31.646576 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/7617925e-d725-4eec-8ea3-7d9eed28d8d7-lib-modules\") pod \"glance-default-internal-api-0\" (UID: \"7617925e-d725-4eec-8ea3-7d9eed28d8d7\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 06 03:57:31 crc kubenswrapper[4980]: I1206 03:57:31.646593 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/7617925e-d725-4eec-8ea3-7d9eed28d8d7-run\") pod \"glance-default-internal-api-0\" (UID: \"7617925e-d725-4eec-8ea3-7d9eed28d8d7\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 06 03:57:31 crc kubenswrapper[4980]: I1206 03:57:31.646639 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/7617925e-d725-4eec-8ea3-7d9eed28d8d7-sys\") pod \"glance-default-internal-api-0\" (UID: \"7617925e-d725-4eec-8ea3-7d9eed28d8d7\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 06 03:57:31 crc kubenswrapper[4980]: I1206 03:57:31.646749 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/7617925e-d725-4eec-8ea3-7d9eed28d8d7-var-locks-brick\") pod \"glance-default-internal-api-0\" (UID: \"7617925e-d725-4eec-8ea3-7d9eed28d8d7\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 06 03:57:31 crc kubenswrapper[4980]: I1206 03:57:31.646886 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7617925e-d725-4eec-8ea3-7d9eed28d8d7-logs\") pod \"glance-default-internal-api-0\" (UID: \"7617925e-d725-4eec-8ea3-7d9eed28d8d7\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 06 03:57:31 crc kubenswrapper[4980]: I1206 03:57:31.661972 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7617925e-d725-4eec-8ea3-7d9eed28d8d7-config-data\") pod \"glance-default-internal-api-0\" (UID: \"7617925e-d725-4eec-8ea3-7d9eed28d8d7\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 06 03:57:31 crc kubenswrapper[4980]: I1206 03:57:31.664766 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f5jx4\" (UniqueName: \"kubernetes.io/projected/7617925e-d725-4eec-8ea3-7d9eed28d8d7-kube-api-access-f5jx4\") pod \"glance-default-internal-api-0\" (UID: \"7617925e-d725-4eec-8ea3-7d9eed28d8d7\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 06 03:57:31 crc kubenswrapper[4980]: I1206 03:57:31.672578 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7617925e-d725-4eec-8ea3-7d9eed28d8d7-scripts\") pod \"glance-default-internal-api-0\" (UID: \"7617925e-d725-4eec-8ea3-7d9eed28d8d7\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 06 03:57:31 crc kubenswrapper[4980]: I1206 03:57:31.691369 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-internal-api-0\" (UID: \"7617925e-d725-4eec-8ea3-7d9eed28d8d7\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 06 03:57:31 crc kubenswrapper[4980]: I1206 03:57:31.700485 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-internal-api-0\" (UID: \"7617925e-d725-4eec-8ea3-7d9eed28d8d7\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 06 03:57:31 crc kubenswrapper[4980]: I1206 03:57:31.926443 4980 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-external-api-0"] Dec 06 03:57:31 crc kubenswrapper[4980]: I1206 03:57:31.981155 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 06 03:57:31 crc kubenswrapper[4980]: I1206 03:57:31.988428 4980 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-0"] Dec 06 03:57:32 crc kubenswrapper[4980]: I1206 03:57:32.513455 4980 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-0"] Dec 06 03:57:32 crc kubenswrapper[4980]: W1206 03:57:32.521338 4980 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7617925e_d725_4eec_8ea3_7d9eed28d8d7.slice/crio-f7478d5ccf115facf349c146b6ecbbb88b9e7497a732a186f3a492139449ee3b WatchSource:0}: Error finding container f7478d5ccf115facf349c146b6ecbbb88b9e7497a732a186f3a492139449ee3b: Status 404 returned error can't find the container with id f7478d5ccf115facf349c146b6ecbbb88b9e7497a732a186f3a492139449ee3b Dec 06 03:57:32 crc kubenswrapper[4980]: I1206 03:57:32.816877 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-0" event={"ID":"7617925e-d725-4eec-8ea3-7d9eed28d8d7","Type":"ContainerStarted","Data":"0f6e411d8fa016a11ea18efdac76970a9ab27210b68737090e8323341f218bca"} Dec 06 03:57:32 crc kubenswrapper[4980]: I1206 03:57:32.816930 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-0" event={"ID":"7617925e-d725-4eec-8ea3-7d9eed28d8d7","Type":"ContainerStarted","Data":"f7478d5ccf115facf349c146b6ecbbb88b9e7497a732a186f3a492139449ee3b"} Dec 06 03:57:32 crc kubenswrapper[4980]: I1206 03:57:32.819159 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-0" event={"ID":"7d66b977-3b53-4fe4-bc08-9e9083b32d76","Type":"ContainerStarted","Data":"68c0b6d46675b06eb4656fb390b7382da33be414c5cbf6039ff75f61949cff58"} Dec 06 03:57:32 crc kubenswrapper[4980]: I1206 03:57:32.819210 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-0" event={"ID":"7d66b977-3b53-4fe4-bc08-9e9083b32d76","Type":"ContainerStarted","Data":"a3b45fd567bca6830803ed66449513fe6d25421e9e02f1227de29ac57265ed76"} Dec 06 03:57:32 crc kubenswrapper[4980]: I1206 03:57:32.819231 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-0" event={"ID":"7d66b977-3b53-4fe4-bc08-9e9083b32d76","Type":"ContainerStarted","Data":"f2a866df9373ad31c1d1f83a161c038b1407b0efda10940c73b0894a91f5a561"} Dec 06 03:57:32 crc kubenswrapper[4980]: I1206 03:57:32.819247 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-0" event={"ID":"7d66b977-3b53-4fe4-bc08-9e9083b32d76","Type":"ContainerStarted","Data":"8d48c17c1cff0953442cd315bf504ff3f5d511ef58312d5b6b7c4510a686edcb"} Dec 06 03:57:32 crc kubenswrapper[4980]: I1206 03:57:32.862502 4980 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/glance-default-external-api-0" podStartSLOduration=1.8624773289999998 podStartE2EDuration="1.862477329s" podCreationTimestamp="2025-12-06 03:57:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 03:57:32.850286649 +0000 UTC m=+1412.095037950" watchObservedRunningTime="2025-12-06 03:57:32.862477329 +0000 UTC m=+1412.107228610" Dec 06 03:57:33 crc kubenswrapper[4980]: I1206 03:57:33.830835 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-0" event={"ID":"7617925e-d725-4eec-8ea3-7d9eed28d8d7","Type":"ContainerStarted","Data":"d05c2cfc257b9632f9aa3799ab3eecabf9d145e8dd561782398290a955926d26"} Dec 06 03:57:33 crc kubenswrapper[4980]: I1206 03:57:33.831622 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-0" event={"ID":"7617925e-d725-4eec-8ea3-7d9eed28d8d7","Type":"ContainerStarted","Data":"bea20c4326b1a79f6f635f2c7739d142f2de487f8f5362eac08d4b294874138f"} Dec 06 03:57:33 crc kubenswrapper[4980]: I1206 03:57:33.831867 4980 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-internal-api-0" podUID="7617925e-d725-4eec-8ea3-7d9eed28d8d7" containerName="glance-log" containerID="cri-o://0f6e411d8fa016a11ea18efdac76970a9ab27210b68737090e8323341f218bca" gracePeriod=30 Dec 06 03:57:33 crc kubenswrapper[4980]: I1206 03:57:33.832051 4980 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-internal-api-0" podUID="7617925e-d725-4eec-8ea3-7d9eed28d8d7" containerName="glance-api" containerID="cri-o://d05c2cfc257b9632f9aa3799ab3eecabf9d145e8dd561782398290a955926d26" gracePeriod=30 Dec 06 03:57:33 crc kubenswrapper[4980]: I1206 03:57:33.832129 4980 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-internal-api-0" podUID="7617925e-d725-4eec-8ea3-7d9eed28d8d7" containerName="glance-httpd" containerID="cri-o://bea20c4326b1a79f6f635f2c7739d142f2de487f8f5362eac08d4b294874138f" gracePeriod=30 Dec 06 03:57:33 crc kubenswrapper[4980]: I1206 03:57:33.859933 4980 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/glance-default-internal-api-0" podStartSLOduration=3.859910612 podStartE2EDuration="3.859910612s" podCreationTimestamp="2025-12-06 03:57:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 03:57:33.857340578 +0000 UTC m=+1413.102091859" watchObservedRunningTime="2025-12-06 03:57:33.859910612 +0000 UTC m=+1413.104661903" Dec 06 03:57:34 crc kubenswrapper[4980]: I1206 03:57:34.351072 4980 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 06 03:57:34 crc kubenswrapper[4980]: I1206 03:57:34.527932 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/7617925e-d725-4eec-8ea3-7d9eed28d8d7-sys\") pod \"7617925e-d725-4eec-8ea3-7d9eed28d8d7\" (UID: \"7617925e-d725-4eec-8ea3-7d9eed28d8d7\") " Dec 06 03:57:34 crc kubenswrapper[4980]: I1206 03:57:34.528218 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7617925e-d725-4eec-8ea3-7d9eed28d8d7-config-data\") pod \"7617925e-d725-4eec-8ea3-7d9eed28d8d7\" (UID: \"7617925e-d725-4eec-8ea3-7d9eed28d8d7\") " Dec 06 03:57:34 crc kubenswrapper[4980]: I1206 03:57:34.528244 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7617925e-d725-4eec-8ea3-7d9eed28d8d7-scripts\") pod \"7617925e-d725-4eec-8ea3-7d9eed28d8d7\" (UID: \"7617925e-d725-4eec-8ea3-7d9eed28d8d7\") " Dec 06 03:57:34 crc kubenswrapper[4980]: I1206 03:57:34.528011 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/7617925e-d725-4eec-8ea3-7d9eed28d8d7-sys" (OuterVolumeSpecName: "sys") pod "7617925e-d725-4eec-8ea3-7d9eed28d8d7" (UID: "7617925e-d725-4eec-8ea3-7d9eed28d8d7"). InnerVolumeSpecName "sys". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 03:57:34 crc kubenswrapper[4980]: I1206 03:57:34.528307 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance-cache\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"7617925e-d725-4eec-8ea3-7d9eed28d8d7\" (UID: \"7617925e-d725-4eec-8ea3-7d9eed28d8d7\") " Dec 06 03:57:34 crc kubenswrapper[4980]: I1206 03:57:34.528381 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/7617925e-d725-4eec-8ea3-7d9eed28d8d7-etc-nvme\") pod \"7617925e-d725-4eec-8ea3-7d9eed28d8d7\" (UID: \"7617925e-d725-4eec-8ea3-7d9eed28d8d7\") " Dec 06 03:57:34 crc kubenswrapper[4980]: I1206 03:57:34.528436 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/7617925e-d725-4eec-8ea3-7d9eed28d8d7-var-locks-brick\") pod \"7617925e-d725-4eec-8ea3-7d9eed28d8d7\" (UID: \"7617925e-d725-4eec-8ea3-7d9eed28d8d7\") " Dec 06 03:57:34 crc kubenswrapper[4980]: I1206 03:57:34.528463 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f5jx4\" (UniqueName: \"kubernetes.io/projected/7617925e-d725-4eec-8ea3-7d9eed28d8d7-kube-api-access-f5jx4\") pod \"7617925e-d725-4eec-8ea3-7d9eed28d8d7\" (UID: \"7617925e-d725-4eec-8ea3-7d9eed28d8d7\") " Dec 06 03:57:34 crc kubenswrapper[4980]: I1206 03:57:34.528503 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/7617925e-d725-4eec-8ea3-7d9eed28d8d7-lib-modules\") pod \"7617925e-d725-4eec-8ea3-7d9eed28d8d7\" (UID: \"7617925e-d725-4eec-8ea3-7d9eed28d8d7\") " Dec 06 03:57:34 crc kubenswrapper[4980]: I1206 03:57:34.528576 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/7617925e-d725-4eec-8ea3-7d9eed28d8d7-run\") pod \"7617925e-d725-4eec-8ea3-7d9eed28d8d7\" (UID: \"7617925e-d725-4eec-8ea3-7d9eed28d8d7\") " Dec 06 03:57:34 crc kubenswrapper[4980]: I1206 03:57:34.528609 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/7617925e-d725-4eec-8ea3-7d9eed28d8d7-etc-iscsi\") pod \"7617925e-d725-4eec-8ea3-7d9eed28d8d7\" (UID: \"7617925e-d725-4eec-8ea3-7d9eed28d8d7\") " Dec 06 03:57:34 crc kubenswrapper[4980]: I1206 03:57:34.528632 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"7617925e-d725-4eec-8ea3-7d9eed28d8d7\" (UID: \"7617925e-d725-4eec-8ea3-7d9eed28d8d7\") " Dec 06 03:57:34 crc kubenswrapper[4980]: I1206 03:57:34.528674 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/7617925e-d725-4eec-8ea3-7d9eed28d8d7-httpd-run\") pod \"7617925e-d725-4eec-8ea3-7d9eed28d8d7\" (UID: \"7617925e-d725-4eec-8ea3-7d9eed28d8d7\") " Dec 06 03:57:34 crc kubenswrapper[4980]: I1206 03:57:34.528695 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/7617925e-d725-4eec-8ea3-7d9eed28d8d7-dev\") pod \"7617925e-d725-4eec-8ea3-7d9eed28d8d7\" (UID: \"7617925e-d725-4eec-8ea3-7d9eed28d8d7\") " Dec 06 03:57:34 crc kubenswrapper[4980]: I1206 03:57:34.528733 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7617925e-d725-4eec-8ea3-7d9eed28d8d7-logs\") pod \"7617925e-d725-4eec-8ea3-7d9eed28d8d7\" (UID: \"7617925e-d725-4eec-8ea3-7d9eed28d8d7\") " Dec 06 03:57:34 crc kubenswrapper[4980]: I1206 03:57:34.528990 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/7617925e-d725-4eec-8ea3-7d9eed28d8d7-lib-modules" (OuterVolumeSpecName: "lib-modules") pod "7617925e-d725-4eec-8ea3-7d9eed28d8d7" (UID: "7617925e-d725-4eec-8ea3-7d9eed28d8d7"). InnerVolumeSpecName "lib-modules". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 03:57:34 crc kubenswrapper[4980]: I1206 03:57:34.529030 4980 reconciler_common.go:293] "Volume detached for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/7617925e-d725-4eec-8ea3-7d9eed28d8d7-sys\") on node \"crc\" DevicePath \"\"" Dec 06 03:57:34 crc kubenswrapper[4980]: I1206 03:57:34.529026 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/7617925e-d725-4eec-8ea3-7d9eed28d8d7-run" (OuterVolumeSpecName: "run") pod "7617925e-d725-4eec-8ea3-7d9eed28d8d7" (UID: "7617925e-d725-4eec-8ea3-7d9eed28d8d7"). InnerVolumeSpecName "run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 03:57:34 crc kubenswrapper[4980]: I1206 03:57:34.529058 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/7617925e-d725-4eec-8ea3-7d9eed28d8d7-etc-nvme" (OuterVolumeSpecName: "etc-nvme") pod "7617925e-d725-4eec-8ea3-7d9eed28d8d7" (UID: "7617925e-d725-4eec-8ea3-7d9eed28d8d7"). InnerVolumeSpecName "etc-nvme". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 03:57:34 crc kubenswrapper[4980]: I1206 03:57:34.529172 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/7617925e-d725-4eec-8ea3-7d9eed28d8d7-dev" (OuterVolumeSpecName: "dev") pod "7617925e-d725-4eec-8ea3-7d9eed28d8d7" (UID: "7617925e-d725-4eec-8ea3-7d9eed28d8d7"). InnerVolumeSpecName "dev". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 03:57:34 crc kubenswrapper[4980]: I1206 03:57:34.529110 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/7617925e-d725-4eec-8ea3-7d9eed28d8d7-etc-iscsi" (OuterVolumeSpecName: "etc-iscsi") pod "7617925e-d725-4eec-8ea3-7d9eed28d8d7" (UID: "7617925e-d725-4eec-8ea3-7d9eed28d8d7"). InnerVolumeSpecName "etc-iscsi". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 03:57:34 crc kubenswrapper[4980]: I1206 03:57:34.529218 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/7617925e-d725-4eec-8ea3-7d9eed28d8d7-var-locks-brick" (OuterVolumeSpecName: "var-locks-brick") pod "7617925e-d725-4eec-8ea3-7d9eed28d8d7" (UID: "7617925e-d725-4eec-8ea3-7d9eed28d8d7"). InnerVolumeSpecName "var-locks-brick". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 03:57:34 crc kubenswrapper[4980]: I1206 03:57:34.529588 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7617925e-d725-4eec-8ea3-7d9eed28d8d7-logs" (OuterVolumeSpecName: "logs") pod "7617925e-d725-4eec-8ea3-7d9eed28d8d7" (UID: "7617925e-d725-4eec-8ea3-7d9eed28d8d7"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 03:57:34 crc kubenswrapper[4980]: I1206 03:57:34.529487 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7617925e-d725-4eec-8ea3-7d9eed28d8d7-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "7617925e-d725-4eec-8ea3-7d9eed28d8d7" (UID: "7617925e-d725-4eec-8ea3-7d9eed28d8d7"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 03:57:34 crc kubenswrapper[4980]: I1206 03:57:34.534032 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage12-crc" (OuterVolumeSpecName: "glance") pod "7617925e-d725-4eec-8ea3-7d9eed28d8d7" (UID: "7617925e-d725-4eec-8ea3-7d9eed28d8d7"). InnerVolumeSpecName "local-storage12-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 06 03:57:34 crc kubenswrapper[4980]: I1206 03:57:34.534687 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7617925e-d725-4eec-8ea3-7d9eed28d8d7-kube-api-access-f5jx4" (OuterVolumeSpecName: "kube-api-access-f5jx4") pod "7617925e-d725-4eec-8ea3-7d9eed28d8d7" (UID: "7617925e-d725-4eec-8ea3-7d9eed28d8d7"). InnerVolumeSpecName "kube-api-access-f5jx4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 03:57:34 crc kubenswrapper[4980]: I1206 03:57:34.536635 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage01-crc" (OuterVolumeSpecName: "glance-cache") pod "7617925e-d725-4eec-8ea3-7d9eed28d8d7" (UID: "7617925e-d725-4eec-8ea3-7d9eed28d8d7"). InnerVolumeSpecName "local-storage01-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 06 03:57:34 crc kubenswrapper[4980]: I1206 03:57:34.538161 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7617925e-d725-4eec-8ea3-7d9eed28d8d7-scripts" (OuterVolumeSpecName: "scripts") pod "7617925e-d725-4eec-8ea3-7d9eed28d8d7" (UID: "7617925e-d725-4eec-8ea3-7d9eed28d8d7"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 03:57:34 crc kubenswrapper[4980]: I1206 03:57:34.631272 4980 reconciler_common.go:293] "Volume detached for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/7617925e-d725-4eec-8ea3-7d9eed28d8d7-etc-nvme\") on node \"crc\" DevicePath \"\"" Dec 06 03:57:34 crc kubenswrapper[4980]: I1206 03:57:34.631322 4980 reconciler_common.go:293] "Volume detached for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/7617925e-d725-4eec-8ea3-7d9eed28d8d7-var-locks-brick\") on node \"crc\" DevicePath \"\"" Dec 06 03:57:34 crc kubenswrapper[4980]: I1206 03:57:34.631342 4980 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f5jx4\" (UniqueName: \"kubernetes.io/projected/7617925e-d725-4eec-8ea3-7d9eed28d8d7-kube-api-access-f5jx4\") on node \"crc\" DevicePath \"\"" Dec 06 03:57:34 crc kubenswrapper[4980]: I1206 03:57:34.631364 4980 reconciler_common.go:293] "Volume detached for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/7617925e-d725-4eec-8ea3-7d9eed28d8d7-lib-modules\") on node \"crc\" DevicePath \"\"" Dec 06 03:57:34 crc kubenswrapper[4980]: I1206 03:57:34.631381 4980 reconciler_common.go:293] "Volume detached for volume \"run\" (UniqueName: \"kubernetes.io/host-path/7617925e-d725-4eec-8ea3-7d9eed28d8d7-run\") on node \"crc\" DevicePath \"\"" Dec 06 03:57:34 crc kubenswrapper[4980]: I1206 03:57:34.631400 4980 reconciler_common.go:293] "Volume detached for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/7617925e-d725-4eec-8ea3-7d9eed28d8d7-etc-iscsi\") on node \"crc\" DevicePath \"\"" Dec 06 03:57:34 crc kubenswrapper[4980]: I1206 03:57:34.631448 4980 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") on node \"crc\" " Dec 06 03:57:34 crc kubenswrapper[4980]: I1206 03:57:34.631465 4980 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/7617925e-d725-4eec-8ea3-7d9eed28d8d7-httpd-run\") on node \"crc\" DevicePath \"\"" Dec 06 03:57:34 crc kubenswrapper[4980]: I1206 03:57:34.631483 4980 reconciler_common.go:293] "Volume detached for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/7617925e-d725-4eec-8ea3-7d9eed28d8d7-dev\") on node \"crc\" DevicePath \"\"" Dec 06 03:57:34 crc kubenswrapper[4980]: I1206 03:57:34.631498 4980 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7617925e-d725-4eec-8ea3-7d9eed28d8d7-logs\") on node \"crc\" DevicePath \"\"" Dec 06 03:57:34 crc kubenswrapper[4980]: I1206 03:57:34.631537 4980 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7617925e-d725-4eec-8ea3-7d9eed28d8d7-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 03:57:34 crc kubenswrapper[4980]: I1206 03:57:34.631564 4980 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") on node \"crc\" " Dec 06 03:57:34 crc kubenswrapper[4980]: I1206 03:57:34.647854 4980 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage12-crc" (UniqueName: "kubernetes.io/local-volume/local-storage12-crc") on node "crc" Dec 06 03:57:34 crc kubenswrapper[4980]: I1206 03:57:34.651041 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7617925e-d725-4eec-8ea3-7d9eed28d8d7-config-data" (OuterVolumeSpecName: "config-data") pod "7617925e-d725-4eec-8ea3-7d9eed28d8d7" (UID: "7617925e-d725-4eec-8ea3-7d9eed28d8d7"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 03:57:34 crc kubenswrapper[4980]: I1206 03:57:34.705591 4980 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage01-crc" (UniqueName: "kubernetes.io/local-volume/local-storage01-crc") on node "crc" Dec 06 03:57:34 crc kubenswrapper[4980]: I1206 03:57:34.733627 4980 reconciler_common.go:293] "Volume detached for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") on node \"crc\" DevicePath \"\"" Dec 06 03:57:34 crc kubenswrapper[4980]: I1206 03:57:34.733685 4980 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7617925e-d725-4eec-8ea3-7d9eed28d8d7-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 03:57:34 crc kubenswrapper[4980]: I1206 03:57:34.733703 4980 reconciler_common.go:293] "Volume detached for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") on node \"crc\" DevicePath \"\"" Dec 06 03:57:34 crc kubenswrapper[4980]: I1206 03:57:34.842065 4980 generic.go:334] "Generic (PLEG): container finished" podID="7617925e-d725-4eec-8ea3-7d9eed28d8d7" containerID="d05c2cfc257b9632f9aa3799ab3eecabf9d145e8dd561782398290a955926d26" exitCode=143 Dec 06 03:57:34 crc kubenswrapper[4980]: I1206 03:57:34.842101 4980 generic.go:334] "Generic (PLEG): container finished" podID="7617925e-d725-4eec-8ea3-7d9eed28d8d7" containerID="bea20c4326b1a79f6f635f2c7739d142f2de487f8f5362eac08d4b294874138f" exitCode=0 Dec 06 03:57:34 crc kubenswrapper[4980]: I1206 03:57:34.842111 4980 generic.go:334] "Generic (PLEG): container finished" podID="7617925e-d725-4eec-8ea3-7d9eed28d8d7" containerID="0f6e411d8fa016a11ea18efdac76970a9ab27210b68737090e8323341f218bca" exitCode=143 Dec 06 03:57:34 crc kubenswrapper[4980]: I1206 03:57:34.842130 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-0" event={"ID":"7617925e-d725-4eec-8ea3-7d9eed28d8d7","Type":"ContainerDied","Data":"d05c2cfc257b9632f9aa3799ab3eecabf9d145e8dd561782398290a955926d26"} Dec 06 03:57:34 crc kubenswrapper[4980]: I1206 03:57:34.842159 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-0" event={"ID":"7617925e-d725-4eec-8ea3-7d9eed28d8d7","Type":"ContainerDied","Data":"bea20c4326b1a79f6f635f2c7739d142f2de487f8f5362eac08d4b294874138f"} Dec 06 03:57:34 crc kubenswrapper[4980]: I1206 03:57:34.842172 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-0" event={"ID":"7617925e-d725-4eec-8ea3-7d9eed28d8d7","Type":"ContainerDied","Data":"0f6e411d8fa016a11ea18efdac76970a9ab27210b68737090e8323341f218bca"} Dec 06 03:57:34 crc kubenswrapper[4980]: I1206 03:57:34.842186 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-0" event={"ID":"7617925e-d725-4eec-8ea3-7d9eed28d8d7","Type":"ContainerDied","Data":"f7478d5ccf115facf349c146b6ecbbb88b9e7497a732a186f3a492139449ee3b"} Dec 06 03:57:34 crc kubenswrapper[4980]: I1206 03:57:34.842206 4980 scope.go:117] "RemoveContainer" containerID="d05c2cfc257b9632f9aa3799ab3eecabf9d145e8dd561782398290a955926d26" Dec 06 03:57:34 crc kubenswrapper[4980]: I1206 03:57:34.842252 4980 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 06 03:57:34 crc kubenswrapper[4980]: I1206 03:57:34.864727 4980 scope.go:117] "RemoveContainer" containerID="bea20c4326b1a79f6f635f2c7739d142f2de487f8f5362eac08d4b294874138f" Dec 06 03:57:34 crc kubenswrapper[4980]: I1206 03:57:34.891774 4980 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-0"] Dec 06 03:57:34 crc kubenswrapper[4980]: I1206 03:57:34.898197 4980 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-0"] Dec 06 03:57:34 crc kubenswrapper[4980]: I1206 03:57:34.917330 4980 scope.go:117] "RemoveContainer" containerID="0f6e411d8fa016a11ea18efdac76970a9ab27210b68737090e8323341f218bca" Dec 06 03:57:34 crc kubenswrapper[4980]: I1206 03:57:34.932527 4980 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-0"] Dec 06 03:57:34 crc kubenswrapper[4980]: E1206 03:57:34.932823 4980 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7617925e-d725-4eec-8ea3-7d9eed28d8d7" containerName="glance-log" Dec 06 03:57:34 crc kubenswrapper[4980]: I1206 03:57:34.932835 4980 state_mem.go:107] "Deleted CPUSet assignment" podUID="7617925e-d725-4eec-8ea3-7d9eed28d8d7" containerName="glance-log" Dec 06 03:57:34 crc kubenswrapper[4980]: E1206 03:57:34.932862 4980 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7617925e-d725-4eec-8ea3-7d9eed28d8d7" containerName="glance-httpd" Dec 06 03:57:34 crc kubenswrapper[4980]: I1206 03:57:34.932870 4980 state_mem.go:107] "Deleted CPUSet assignment" podUID="7617925e-d725-4eec-8ea3-7d9eed28d8d7" containerName="glance-httpd" Dec 06 03:57:34 crc kubenswrapper[4980]: E1206 03:57:34.932882 4980 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7617925e-d725-4eec-8ea3-7d9eed28d8d7" containerName="glance-api" Dec 06 03:57:34 crc kubenswrapper[4980]: I1206 03:57:34.932888 4980 state_mem.go:107] "Deleted CPUSet assignment" podUID="7617925e-d725-4eec-8ea3-7d9eed28d8d7" containerName="glance-api" Dec 06 03:57:34 crc kubenswrapper[4980]: I1206 03:57:34.933036 4980 memory_manager.go:354] "RemoveStaleState removing state" podUID="7617925e-d725-4eec-8ea3-7d9eed28d8d7" containerName="glance-httpd" Dec 06 03:57:34 crc kubenswrapper[4980]: I1206 03:57:34.933053 4980 memory_manager.go:354] "RemoveStaleState removing state" podUID="7617925e-d725-4eec-8ea3-7d9eed28d8d7" containerName="glance-api" Dec 06 03:57:34 crc kubenswrapper[4980]: I1206 03:57:34.933068 4980 memory_manager.go:354] "RemoveStaleState removing state" podUID="7617925e-d725-4eec-8ea3-7d9eed28d8d7" containerName="glance-log" Dec 06 03:57:34 crc kubenswrapper[4980]: I1206 03:57:34.934111 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 06 03:57:34 crc kubenswrapper[4980]: I1206 03:57:34.936138 4980 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-default-internal-config-data" Dec 06 03:57:34 crc kubenswrapper[4980]: I1206 03:57:34.954688 4980 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-0"] Dec 06 03:57:34 crc kubenswrapper[4980]: I1206 03:57:34.956568 4980 scope.go:117] "RemoveContainer" containerID="d05c2cfc257b9632f9aa3799ab3eecabf9d145e8dd561782398290a955926d26" Dec 06 03:57:34 crc kubenswrapper[4980]: E1206 03:57:34.957035 4980 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d05c2cfc257b9632f9aa3799ab3eecabf9d145e8dd561782398290a955926d26\": container with ID starting with d05c2cfc257b9632f9aa3799ab3eecabf9d145e8dd561782398290a955926d26 not found: ID does not exist" containerID="d05c2cfc257b9632f9aa3799ab3eecabf9d145e8dd561782398290a955926d26" Dec 06 03:57:34 crc kubenswrapper[4980]: I1206 03:57:34.957066 4980 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d05c2cfc257b9632f9aa3799ab3eecabf9d145e8dd561782398290a955926d26"} err="failed to get container status \"d05c2cfc257b9632f9aa3799ab3eecabf9d145e8dd561782398290a955926d26\": rpc error: code = NotFound desc = could not find container \"d05c2cfc257b9632f9aa3799ab3eecabf9d145e8dd561782398290a955926d26\": container with ID starting with d05c2cfc257b9632f9aa3799ab3eecabf9d145e8dd561782398290a955926d26 not found: ID does not exist" Dec 06 03:57:34 crc kubenswrapper[4980]: I1206 03:57:34.957093 4980 scope.go:117] "RemoveContainer" containerID="bea20c4326b1a79f6f635f2c7739d142f2de487f8f5362eac08d4b294874138f" Dec 06 03:57:34 crc kubenswrapper[4980]: E1206 03:57:34.957701 4980 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bea20c4326b1a79f6f635f2c7739d142f2de487f8f5362eac08d4b294874138f\": container with ID starting with bea20c4326b1a79f6f635f2c7739d142f2de487f8f5362eac08d4b294874138f not found: ID does not exist" containerID="bea20c4326b1a79f6f635f2c7739d142f2de487f8f5362eac08d4b294874138f" Dec 06 03:57:34 crc kubenswrapper[4980]: I1206 03:57:34.957729 4980 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bea20c4326b1a79f6f635f2c7739d142f2de487f8f5362eac08d4b294874138f"} err="failed to get container status \"bea20c4326b1a79f6f635f2c7739d142f2de487f8f5362eac08d4b294874138f\": rpc error: code = NotFound desc = could not find container \"bea20c4326b1a79f6f635f2c7739d142f2de487f8f5362eac08d4b294874138f\": container with ID starting with bea20c4326b1a79f6f635f2c7739d142f2de487f8f5362eac08d4b294874138f not found: ID does not exist" Dec 06 03:57:34 crc kubenswrapper[4980]: I1206 03:57:34.957747 4980 scope.go:117] "RemoveContainer" containerID="0f6e411d8fa016a11ea18efdac76970a9ab27210b68737090e8323341f218bca" Dec 06 03:57:34 crc kubenswrapper[4980]: E1206 03:57:34.958023 4980 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0f6e411d8fa016a11ea18efdac76970a9ab27210b68737090e8323341f218bca\": container with ID starting with 0f6e411d8fa016a11ea18efdac76970a9ab27210b68737090e8323341f218bca not found: ID does not exist" containerID="0f6e411d8fa016a11ea18efdac76970a9ab27210b68737090e8323341f218bca" Dec 06 03:57:34 crc kubenswrapper[4980]: I1206 03:57:34.958056 4980 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0f6e411d8fa016a11ea18efdac76970a9ab27210b68737090e8323341f218bca"} err="failed to get container status \"0f6e411d8fa016a11ea18efdac76970a9ab27210b68737090e8323341f218bca\": rpc error: code = NotFound desc = could not find container \"0f6e411d8fa016a11ea18efdac76970a9ab27210b68737090e8323341f218bca\": container with ID starting with 0f6e411d8fa016a11ea18efdac76970a9ab27210b68737090e8323341f218bca not found: ID does not exist" Dec 06 03:57:34 crc kubenswrapper[4980]: I1206 03:57:34.958098 4980 scope.go:117] "RemoveContainer" containerID="d05c2cfc257b9632f9aa3799ab3eecabf9d145e8dd561782398290a955926d26" Dec 06 03:57:34 crc kubenswrapper[4980]: I1206 03:57:34.958378 4980 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d05c2cfc257b9632f9aa3799ab3eecabf9d145e8dd561782398290a955926d26"} err="failed to get container status \"d05c2cfc257b9632f9aa3799ab3eecabf9d145e8dd561782398290a955926d26\": rpc error: code = NotFound desc = could not find container \"d05c2cfc257b9632f9aa3799ab3eecabf9d145e8dd561782398290a955926d26\": container with ID starting with d05c2cfc257b9632f9aa3799ab3eecabf9d145e8dd561782398290a955926d26 not found: ID does not exist" Dec 06 03:57:34 crc kubenswrapper[4980]: I1206 03:57:34.958401 4980 scope.go:117] "RemoveContainer" containerID="bea20c4326b1a79f6f635f2c7739d142f2de487f8f5362eac08d4b294874138f" Dec 06 03:57:34 crc kubenswrapper[4980]: I1206 03:57:34.958806 4980 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bea20c4326b1a79f6f635f2c7739d142f2de487f8f5362eac08d4b294874138f"} err="failed to get container status \"bea20c4326b1a79f6f635f2c7739d142f2de487f8f5362eac08d4b294874138f\": rpc error: code = NotFound desc = could not find container \"bea20c4326b1a79f6f635f2c7739d142f2de487f8f5362eac08d4b294874138f\": container with ID starting with bea20c4326b1a79f6f635f2c7739d142f2de487f8f5362eac08d4b294874138f not found: ID does not exist" Dec 06 03:57:34 crc kubenswrapper[4980]: I1206 03:57:34.958825 4980 scope.go:117] "RemoveContainer" containerID="0f6e411d8fa016a11ea18efdac76970a9ab27210b68737090e8323341f218bca" Dec 06 03:57:34 crc kubenswrapper[4980]: I1206 03:57:34.966216 4980 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0f6e411d8fa016a11ea18efdac76970a9ab27210b68737090e8323341f218bca"} err="failed to get container status \"0f6e411d8fa016a11ea18efdac76970a9ab27210b68737090e8323341f218bca\": rpc error: code = NotFound desc = could not find container \"0f6e411d8fa016a11ea18efdac76970a9ab27210b68737090e8323341f218bca\": container with ID starting with 0f6e411d8fa016a11ea18efdac76970a9ab27210b68737090e8323341f218bca not found: ID does not exist" Dec 06 03:57:34 crc kubenswrapper[4980]: I1206 03:57:34.966264 4980 scope.go:117] "RemoveContainer" containerID="d05c2cfc257b9632f9aa3799ab3eecabf9d145e8dd561782398290a955926d26" Dec 06 03:57:34 crc kubenswrapper[4980]: I1206 03:57:34.966751 4980 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d05c2cfc257b9632f9aa3799ab3eecabf9d145e8dd561782398290a955926d26"} err="failed to get container status \"d05c2cfc257b9632f9aa3799ab3eecabf9d145e8dd561782398290a955926d26\": rpc error: code = NotFound desc = could not find container \"d05c2cfc257b9632f9aa3799ab3eecabf9d145e8dd561782398290a955926d26\": container with ID starting with d05c2cfc257b9632f9aa3799ab3eecabf9d145e8dd561782398290a955926d26 not found: ID does not exist" Dec 06 03:57:34 crc kubenswrapper[4980]: I1206 03:57:34.966829 4980 scope.go:117] "RemoveContainer" containerID="bea20c4326b1a79f6f635f2c7739d142f2de487f8f5362eac08d4b294874138f" Dec 06 03:57:34 crc kubenswrapper[4980]: I1206 03:57:34.967271 4980 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bea20c4326b1a79f6f635f2c7739d142f2de487f8f5362eac08d4b294874138f"} err="failed to get container status \"bea20c4326b1a79f6f635f2c7739d142f2de487f8f5362eac08d4b294874138f\": rpc error: code = NotFound desc = could not find container \"bea20c4326b1a79f6f635f2c7739d142f2de487f8f5362eac08d4b294874138f\": container with ID starting with bea20c4326b1a79f6f635f2c7739d142f2de487f8f5362eac08d4b294874138f not found: ID does not exist" Dec 06 03:57:34 crc kubenswrapper[4980]: I1206 03:57:34.967300 4980 scope.go:117] "RemoveContainer" containerID="0f6e411d8fa016a11ea18efdac76970a9ab27210b68737090e8323341f218bca" Dec 06 03:57:34 crc kubenswrapper[4980]: I1206 03:57:34.967645 4980 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0f6e411d8fa016a11ea18efdac76970a9ab27210b68737090e8323341f218bca"} err="failed to get container status \"0f6e411d8fa016a11ea18efdac76970a9ab27210b68737090e8323341f218bca\": rpc error: code = NotFound desc = could not find container \"0f6e411d8fa016a11ea18efdac76970a9ab27210b68737090e8323341f218bca\": container with ID starting with 0f6e411d8fa016a11ea18efdac76970a9ab27210b68737090e8323341f218bca not found: ID does not exist" Dec 06 03:57:35 crc kubenswrapper[4980]: I1206 03:57:35.036760 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jkjth\" (UniqueName: \"kubernetes.io/projected/3ef73fa3-e8eb-4b78-9831-7e3e0f24bdb8-kube-api-access-jkjth\") pod \"glance-default-internal-api-0\" (UID: \"3ef73fa3-e8eb-4b78-9831-7e3e0f24bdb8\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 06 03:57:35 crc kubenswrapper[4980]: I1206 03:57:35.036811 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/3ef73fa3-e8eb-4b78-9831-7e3e0f24bdb8-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"3ef73fa3-e8eb-4b78-9831-7e3e0f24bdb8\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 06 03:57:35 crc kubenswrapper[4980]: I1206 03:57:35.036941 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/3ef73fa3-e8eb-4b78-9831-7e3e0f24bdb8-lib-modules\") pod \"glance-default-internal-api-0\" (UID: \"3ef73fa3-e8eb-4b78-9831-7e3e0f24bdb8\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 06 03:57:35 crc kubenswrapper[4980]: I1206 03:57:35.037043 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3ef73fa3-e8eb-4b78-9831-7e3e0f24bdb8-scripts\") pod \"glance-default-internal-api-0\" (UID: \"3ef73fa3-e8eb-4b78-9831-7e3e0f24bdb8\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 06 03:57:35 crc kubenswrapper[4980]: I1206 03:57:35.037164 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/3ef73fa3-e8eb-4b78-9831-7e3e0f24bdb8-var-locks-brick\") pod \"glance-default-internal-api-0\" (UID: \"3ef73fa3-e8eb-4b78-9831-7e3e0f24bdb8\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 06 03:57:35 crc kubenswrapper[4980]: I1206 03:57:35.037201 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/3ef73fa3-e8eb-4b78-9831-7e3e0f24bdb8-run\") pod \"glance-default-internal-api-0\" (UID: \"3ef73fa3-e8eb-4b78-9831-7e3e0f24bdb8\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 06 03:57:35 crc kubenswrapper[4980]: I1206 03:57:35.037247 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-internal-api-0\" (UID: \"3ef73fa3-e8eb-4b78-9831-7e3e0f24bdb8\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 06 03:57:35 crc kubenswrapper[4980]: I1206 03:57:35.037319 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3ef73fa3-e8eb-4b78-9831-7e3e0f24bdb8-logs\") pod \"glance-default-internal-api-0\" (UID: \"3ef73fa3-e8eb-4b78-9831-7e3e0f24bdb8\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 06 03:57:35 crc kubenswrapper[4980]: I1206 03:57:35.037363 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/3ef73fa3-e8eb-4b78-9831-7e3e0f24bdb8-etc-iscsi\") pod \"glance-default-internal-api-0\" (UID: \"3ef73fa3-e8eb-4b78-9831-7e3e0f24bdb8\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 06 03:57:35 crc kubenswrapper[4980]: I1206 03:57:35.037408 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-internal-api-0\" (UID: \"3ef73fa3-e8eb-4b78-9831-7e3e0f24bdb8\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 06 03:57:35 crc kubenswrapper[4980]: I1206 03:57:35.037494 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/3ef73fa3-e8eb-4b78-9831-7e3e0f24bdb8-sys\") pod \"glance-default-internal-api-0\" (UID: \"3ef73fa3-e8eb-4b78-9831-7e3e0f24bdb8\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 06 03:57:35 crc kubenswrapper[4980]: I1206 03:57:35.037530 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/3ef73fa3-e8eb-4b78-9831-7e3e0f24bdb8-etc-nvme\") pod \"glance-default-internal-api-0\" (UID: \"3ef73fa3-e8eb-4b78-9831-7e3e0f24bdb8\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 06 03:57:35 crc kubenswrapper[4980]: I1206 03:57:35.037567 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3ef73fa3-e8eb-4b78-9831-7e3e0f24bdb8-config-data\") pod \"glance-default-internal-api-0\" (UID: \"3ef73fa3-e8eb-4b78-9831-7e3e0f24bdb8\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 06 03:57:35 crc kubenswrapper[4980]: I1206 03:57:35.037583 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/3ef73fa3-e8eb-4b78-9831-7e3e0f24bdb8-dev\") pod \"glance-default-internal-api-0\" (UID: \"3ef73fa3-e8eb-4b78-9831-7e3e0f24bdb8\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 06 03:57:35 crc kubenswrapper[4980]: I1206 03:57:35.139350 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3ef73fa3-e8eb-4b78-9831-7e3e0f24bdb8-config-data\") pod \"glance-default-internal-api-0\" (UID: \"3ef73fa3-e8eb-4b78-9831-7e3e0f24bdb8\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 06 03:57:35 crc kubenswrapper[4980]: I1206 03:57:35.139397 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/3ef73fa3-e8eb-4b78-9831-7e3e0f24bdb8-dev\") pod \"glance-default-internal-api-0\" (UID: \"3ef73fa3-e8eb-4b78-9831-7e3e0f24bdb8\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 06 03:57:35 crc kubenswrapper[4980]: I1206 03:57:35.139439 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jkjth\" (UniqueName: \"kubernetes.io/projected/3ef73fa3-e8eb-4b78-9831-7e3e0f24bdb8-kube-api-access-jkjth\") pod \"glance-default-internal-api-0\" (UID: \"3ef73fa3-e8eb-4b78-9831-7e3e0f24bdb8\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 06 03:57:35 crc kubenswrapper[4980]: I1206 03:57:35.139462 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/3ef73fa3-e8eb-4b78-9831-7e3e0f24bdb8-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"3ef73fa3-e8eb-4b78-9831-7e3e0f24bdb8\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 06 03:57:35 crc kubenswrapper[4980]: I1206 03:57:35.139493 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/3ef73fa3-e8eb-4b78-9831-7e3e0f24bdb8-lib-modules\") pod \"glance-default-internal-api-0\" (UID: \"3ef73fa3-e8eb-4b78-9831-7e3e0f24bdb8\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 06 03:57:35 crc kubenswrapper[4980]: I1206 03:57:35.139541 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3ef73fa3-e8eb-4b78-9831-7e3e0f24bdb8-scripts\") pod \"glance-default-internal-api-0\" (UID: \"3ef73fa3-e8eb-4b78-9831-7e3e0f24bdb8\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 06 03:57:35 crc kubenswrapper[4980]: I1206 03:57:35.139549 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/3ef73fa3-e8eb-4b78-9831-7e3e0f24bdb8-dev\") pod \"glance-default-internal-api-0\" (UID: \"3ef73fa3-e8eb-4b78-9831-7e3e0f24bdb8\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 06 03:57:35 crc kubenswrapper[4980]: I1206 03:57:35.139589 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/3ef73fa3-e8eb-4b78-9831-7e3e0f24bdb8-var-locks-brick\") pod \"glance-default-internal-api-0\" (UID: \"3ef73fa3-e8eb-4b78-9831-7e3e0f24bdb8\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 06 03:57:35 crc kubenswrapper[4980]: I1206 03:57:35.139613 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/3ef73fa3-e8eb-4b78-9831-7e3e0f24bdb8-run\") pod \"glance-default-internal-api-0\" (UID: \"3ef73fa3-e8eb-4b78-9831-7e3e0f24bdb8\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 06 03:57:35 crc kubenswrapper[4980]: I1206 03:57:35.139616 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/3ef73fa3-e8eb-4b78-9831-7e3e0f24bdb8-lib-modules\") pod \"glance-default-internal-api-0\" (UID: \"3ef73fa3-e8eb-4b78-9831-7e3e0f24bdb8\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 06 03:57:35 crc kubenswrapper[4980]: I1206 03:57:35.139641 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-internal-api-0\" (UID: \"3ef73fa3-e8eb-4b78-9831-7e3e0f24bdb8\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 06 03:57:35 crc kubenswrapper[4980]: I1206 03:57:35.139673 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3ef73fa3-e8eb-4b78-9831-7e3e0f24bdb8-logs\") pod \"glance-default-internal-api-0\" (UID: \"3ef73fa3-e8eb-4b78-9831-7e3e0f24bdb8\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 06 03:57:35 crc kubenswrapper[4980]: I1206 03:57:35.139702 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/3ef73fa3-e8eb-4b78-9831-7e3e0f24bdb8-etc-iscsi\") pod \"glance-default-internal-api-0\" (UID: \"3ef73fa3-e8eb-4b78-9831-7e3e0f24bdb8\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 06 03:57:35 crc kubenswrapper[4980]: I1206 03:57:35.139728 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-internal-api-0\" (UID: \"3ef73fa3-e8eb-4b78-9831-7e3e0f24bdb8\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 06 03:57:35 crc kubenswrapper[4980]: I1206 03:57:35.139759 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/3ef73fa3-e8eb-4b78-9831-7e3e0f24bdb8-sys\") pod \"glance-default-internal-api-0\" (UID: \"3ef73fa3-e8eb-4b78-9831-7e3e0f24bdb8\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 06 03:57:35 crc kubenswrapper[4980]: I1206 03:57:35.139778 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/3ef73fa3-e8eb-4b78-9831-7e3e0f24bdb8-etc-nvme\") pod \"glance-default-internal-api-0\" (UID: \"3ef73fa3-e8eb-4b78-9831-7e3e0f24bdb8\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 06 03:57:35 crc kubenswrapper[4980]: I1206 03:57:35.139871 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/3ef73fa3-e8eb-4b78-9831-7e3e0f24bdb8-etc-nvme\") pod \"glance-default-internal-api-0\" (UID: \"3ef73fa3-e8eb-4b78-9831-7e3e0f24bdb8\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 06 03:57:35 crc kubenswrapper[4980]: I1206 03:57:35.139908 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/3ef73fa3-e8eb-4b78-9831-7e3e0f24bdb8-etc-iscsi\") pod \"glance-default-internal-api-0\" (UID: \"3ef73fa3-e8eb-4b78-9831-7e3e0f24bdb8\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 06 03:57:35 crc kubenswrapper[4980]: I1206 03:57:35.139993 4980 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-internal-api-0\" (UID: \"3ef73fa3-e8eb-4b78-9831-7e3e0f24bdb8\") device mount path \"/mnt/openstack/pv01\"" pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 06 03:57:35 crc kubenswrapper[4980]: I1206 03:57:35.140146 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/3ef73fa3-e8eb-4b78-9831-7e3e0f24bdb8-run\") pod \"glance-default-internal-api-0\" (UID: \"3ef73fa3-e8eb-4b78-9831-7e3e0f24bdb8\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 06 03:57:35 crc kubenswrapper[4980]: I1206 03:57:35.140212 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3ef73fa3-e8eb-4b78-9831-7e3e0f24bdb8-logs\") pod \"glance-default-internal-api-0\" (UID: \"3ef73fa3-e8eb-4b78-9831-7e3e0f24bdb8\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 06 03:57:35 crc kubenswrapper[4980]: I1206 03:57:35.140214 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/3ef73fa3-e8eb-4b78-9831-7e3e0f24bdb8-sys\") pod \"glance-default-internal-api-0\" (UID: \"3ef73fa3-e8eb-4b78-9831-7e3e0f24bdb8\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 06 03:57:35 crc kubenswrapper[4980]: I1206 03:57:35.140264 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/3ef73fa3-e8eb-4b78-9831-7e3e0f24bdb8-var-locks-brick\") pod \"glance-default-internal-api-0\" (UID: \"3ef73fa3-e8eb-4b78-9831-7e3e0f24bdb8\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 06 03:57:35 crc kubenswrapper[4980]: I1206 03:57:35.140303 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/3ef73fa3-e8eb-4b78-9831-7e3e0f24bdb8-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"3ef73fa3-e8eb-4b78-9831-7e3e0f24bdb8\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 06 03:57:35 crc kubenswrapper[4980]: I1206 03:57:35.140317 4980 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-internal-api-0\" (UID: \"3ef73fa3-e8eb-4b78-9831-7e3e0f24bdb8\") device mount path \"/mnt/openstack/pv12\"" pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 06 03:57:35 crc kubenswrapper[4980]: I1206 03:57:35.143928 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3ef73fa3-e8eb-4b78-9831-7e3e0f24bdb8-config-data\") pod \"glance-default-internal-api-0\" (UID: \"3ef73fa3-e8eb-4b78-9831-7e3e0f24bdb8\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 06 03:57:35 crc kubenswrapper[4980]: I1206 03:57:35.144430 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3ef73fa3-e8eb-4b78-9831-7e3e0f24bdb8-scripts\") pod \"glance-default-internal-api-0\" (UID: \"3ef73fa3-e8eb-4b78-9831-7e3e0f24bdb8\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 06 03:57:35 crc kubenswrapper[4980]: I1206 03:57:35.159883 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-internal-api-0\" (UID: \"3ef73fa3-e8eb-4b78-9831-7e3e0f24bdb8\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 06 03:57:35 crc kubenswrapper[4980]: I1206 03:57:35.166751 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-internal-api-0\" (UID: \"3ef73fa3-e8eb-4b78-9831-7e3e0f24bdb8\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 06 03:57:35 crc kubenswrapper[4980]: I1206 03:57:35.167540 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jkjth\" (UniqueName: \"kubernetes.io/projected/3ef73fa3-e8eb-4b78-9831-7e3e0f24bdb8-kube-api-access-jkjth\") pod \"glance-default-internal-api-0\" (UID: \"3ef73fa3-e8eb-4b78-9831-7e3e0f24bdb8\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 06 03:57:35 crc kubenswrapper[4980]: I1206 03:57:35.198367 4980 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7617925e-d725-4eec-8ea3-7d9eed28d8d7" path="/var/lib/kubelet/pods/7617925e-d725-4eec-8ea3-7d9eed28d8d7/volumes" Dec 06 03:57:35 crc kubenswrapper[4980]: I1206 03:57:35.260160 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 06 03:57:35 crc kubenswrapper[4980]: I1206 03:57:35.757774 4980 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-0"] Dec 06 03:57:35 crc kubenswrapper[4980]: I1206 03:57:35.854689 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-0" event={"ID":"3ef73fa3-e8eb-4b78-9831-7e3e0f24bdb8","Type":"ContainerStarted","Data":"95e24dced4a6fa6d657030e1b26201f7dab2392b15ea2926ac35be4658c8767d"} Dec 06 03:57:36 crc kubenswrapper[4980]: I1206 03:57:36.874497 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-0" event={"ID":"3ef73fa3-e8eb-4b78-9831-7e3e0f24bdb8","Type":"ContainerStarted","Data":"480a3a396cafb96c695b846ad3ba450303ee985d2eb24f5fec3d66a770a6b814"} Dec 06 03:57:36 crc kubenswrapper[4980]: I1206 03:57:36.875364 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-0" event={"ID":"3ef73fa3-e8eb-4b78-9831-7e3e0f24bdb8","Type":"ContainerStarted","Data":"d242deb521435afec36a4f76e01c9b146085b07837d657378db253b5857e1bde"} Dec 06 03:57:36 crc kubenswrapper[4980]: I1206 03:57:36.875398 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-0" event={"ID":"3ef73fa3-e8eb-4b78-9831-7e3e0f24bdb8","Type":"ContainerStarted","Data":"772a71fa70b20cc2e08197e2c594baaa1da7d9953b5e02701dfbe2f61747c35a"} Dec 06 03:57:36 crc kubenswrapper[4980]: I1206 03:57:36.916171 4980 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/glance-default-internal-api-0" podStartSLOduration=2.916134026 podStartE2EDuration="2.916134026s" podCreationTimestamp="2025-12-06 03:57:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 03:57:36.910667479 +0000 UTC m=+1416.155418750" watchObservedRunningTime="2025-12-06 03:57:36.916134026 +0000 UTC m=+1416.160885307" Dec 06 03:57:41 crc kubenswrapper[4980]: I1206 03:57:41.487237 4980 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-external-api-0" Dec 06 03:57:41 crc kubenswrapper[4980]: I1206 03:57:41.487808 4980 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-external-api-0" Dec 06 03:57:41 crc kubenswrapper[4980]: I1206 03:57:41.487819 4980 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-external-api-0" Dec 06 03:57:41 crc kubenswrapper[4980]: I1206 03:57:41.512757 4980 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-external-api-0" Dec 06 03:57:41 crc kubenswrapper[4980]: I1206 03:57:41.526749 4980 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-external-api-0" Dec 06 03:57:41 crc kubenswrapper[4980]: I1206 03:57:41.610592 4980 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-external-api-0" Dec 06 03:57:41 crc kubenswrapper[4980]: I1206 03:57:41.922214 4980 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-external-api-0" Dec 06 03:57:41 crc kubenswrapper[4980]: I1206 03:57:41.922284 4980 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-external-api-0" Dec 06 03:57:41 crc kubenswrapper[4980]: I1206 03:57:41.922307 4980 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-external-api-0" Dec 06 03:57:41 crc kubenswrapper[4980]: I1206 03:57:41.944540 4980 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-external-api-0" Dec 06 03:57:41 crc kubenswrapper[4980]: I1206 03:57:41.945563 4980 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-external-api-0" Dec 06 03:57:41 crc kubenswrapper[4980]: I1206 03:57:41.951548 4980 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-external-api-0" Dec 06 03:57:45 crc kubenswrapper[4980]: I1206 03:57:45.260641 4980 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 06 03:57:45 crc kubenswrapper[4980]: I1206 03:57:45.261448 4980 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 06 03:57:45 crc kubenswrapper[4980]: I1206 03:57:45.261479 4980 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 06 03:57:45 crc kubenswrapper[4980]: I1206 03:57:45.298857 4980 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 06 03:57:45 crc kubenswrapper[4980]: I1206 03:57:45.311626 4980 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 06 03:57:45 crc kubenswrapper[4980]: I1206 03:57:45.336570 4980 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 06 03:57:45 crc kubenswrapper[4980]: I1206 03:57:45.963302 4980 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 06 03:57:45 crc kubenswrapper[4980]: I1206 03:57:45.966744 4980 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 06 03:57:45 crc kubenswrapper[4980]: I1206 03:57:45.966760 4980 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 06 03:57:45 crc kubenswrapper[4980]: I1206 03:57:45.976266 4980 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 06 03:57:45 crc kubenswrapper[4980]: I1206 03:57:45.986990 4980 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 06 03:57:45 crc kubenswrapper[4980]: I1206 03:57:45.990096 4980 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 06 03:58:25 crc kubenswrapper[4980]: I1206 03:58:25.138158 4980 patch_prober.go:28] interesting pod/machine-config-daemon-r5zfb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 03:58:25 crc kubenswrapper[4980]: I1206 03:58:25.138882 4980 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-r5zfb" podUID="320f44d1-a671-4a91-b328-a8b0fdd8f23a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 03:58:55 crc kubenswrapper[4980]: I1206 03:58:55.137768 4980 patch_prober.go:28] interesting pod/machine-config-daemon-r5zfb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 03:58:55 crc kubenswrapper[4980]: I1206 03:58:55.138339 4980 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-r5zfb" podUID="320f44d1-a671-4a91-b328-a8b0fdd8f23a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 03:59:25 crc kubenswrapper[4980]: I1206 03:59:25.137306 4980 patch_prober.go:28] interesting pod/machine-config-daemon-r5zfb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 03:59:25 crc kubenswrapper[4980]: I1206 03:59:25.138052 4980 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-r5zfb" podUID="320f44d1-a671-4a91-b328-a8b0fdd8f23a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 03:59:25 crc kubenswrapper[4980]: I1206 03:59:25.138138 4980 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-r5zfb" Dec 06 03:59:25 crc kubenswrapper[4980]: I1206 03:59:25.139246 4980 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"e8878e13c398cceec46bb52c4c9ea0e06dbf4317e6ad39b69ca763a0d8b15497"} pod="openshift-machine-config-operator/machine-config-daemon-r5zfb" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 06 03:59:25 crc kubenswrapper[4980]: I1206 03:59:25.139426 4980 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-r5zfb" podUID="320f44d1-a671-4a91-b328-a8b0fdd8f23a" containerName="machine-config-daemon" containerID="cri-o://e8878e13c398cceec46bb52c4c9ea0e06dbf4317e6ad39b69ca763a0d8b15497" gracePeriod=600 Dec 06 03:59:25 crc kubenswrapper[4980]: E1206 03:59:25.267032 4980 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-r5zfb_openshift-machine-config-operator(320f44d1-a671-4a91-b328-a8b0fdd8f23a)\"" pod="openshift-machine-config-operator/machine-config-daemon-r5zfb" podUID="320f44d1-a671-4a91-b328-a8b0fdd8f23a" Dec 06 03:59:25 crc kubenswrapper[4980]: I1206 03:59:25.971005 4980 generic.go:334] "Generic (PLEG): container finished" podID="320f44d1-a671-4a91-b328-a8b0fdd8f23a" containerID="e8878e13c398cceec46bb52c4c9ea0e06dbf4317e6ad39b69ca763a0d8b15497" exitCode=0 Dec 06 03:59:25 crc kubenswrapper[4980]: I1206 03:59:25.971059 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-r5zfb" event={"ID":"320f44d1-a671-4a91-b328-a8b0fdd8f23a","Type":"ContainerDied","Data":"e8878e13c398cceec46bb52c4c9ea0e06dbf4317e6ad39b69ca763a0d8b15497"} Dec 06 03:59:25 crc kubenswrapper[4980]: I1206 03:59:25.971177 4980 scope.go:117] "RemoveContainer" containerID="ed3956e825d000e94b97025c20ede41bab44553dd92465ad73c08a9ec44729eb" Dec 06 03:59:25 crc kubenswrapper[4980]: I1206 03:59:25.971883 4980 scope.go:117] "RemoveContainer" containerID="e8878e13c398cceec46bb52c4c9ea0e06dbf4317e6ad39b69ca763a0d8b15497" Dec 06 03:59:25 crc kubenswrapper[4980]: E1206 03:59:25.972306 4980 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-r5zfb_openshift-machine-config-operator(320f44d1-a671-4a91-b328-a8b0fdd8f23a)\"" pod="openshift-machine-config-operator/machine-config-daemon-r5zfb" podUID="320f44d1-a671-4a91-b328-a8b0fdd8f23a" Dec 06 03:59:38 crc kubenswrapper[4980]: I1206 03:59:38.185044 4980 scope.go:117] "RemoveContainer" containerID="e8878e13c398cceec46bb52c4c9ea0e06dbf4317e6ad39b69ca763a0d8b15497" Dec 06 03:59:38 crc kubenswrapper[4980]: E1206 03:59:38.186012 4980 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-r5zfb_openshift-machine-config-operator(320f44d1-a671-4a91-b328-a8b0fdd8f23a)\"" pod="openshift-machine-config-operator/machine-config-daemon-r5zfb" podUID="320f44d1-a671-4a91-b328-a8b0fdd8f23a" Dec 06 03:59:53 crc kubenswrapper[4980]: I1206 03:59:53.183990 4980 scope.go:117] "RemoveContainer" containerID="e8878e13c398cceec46bb52c4c9ea0e06dbf4317e6ad39b69ca763a0d8b15497" Dec 06 03:59:53 crc kubenswrapper[4980]: E1206 03:59:53.184808 4980 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-r5zfb_openshift-machine-config-operator(320f44d1-a671-4a91-b328-a8b0fdd8f23a)\"" pod="openshift-machine-config-operator/machine-config-daemon-r5zfb" podUID="320f44d1-a671-4a91-b328-a8b0fdd8f23a" Dec 06 04:00:00 crc kubenswrapper[4980]: I1206 04:00:00.190684 4980 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-cache-glance-default-internal-api-0-cleaner-294165644rpt"] Dec 06 04:00:00 crc kubenswrapper[4980]: I1206 04:00:00.192524 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-cache-glance-default-internal-api-0-cleaner-294165644rpt" Dec 06 04:00:00 crc kubenswrapper[4980]: I1206 04:00:00.196406 4980 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29416560-swv82"] Dec 06 04:00:00 crc kubenswrapper[4980]: I1206 04:00:00.198397 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29416560-swv82" Dec 06 04:00:00 crc kubenswrapper[4980]: I1206 04:00:00.201831 4980 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29416560-swv82"] Dec 06 04:00:00 crc kubenswrapper[4980]: I1206 04:00:00.201927 4980 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 06 04:00:00 crc kubenswrapper[4980]: I1206 04:00:00.201975 4980 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 06 04:00:00 crc kubenswrapper[4980]: I1206 04:00:00.209762 4980 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-cache-glance-default-internal-api-0-cleaner-294165644rpt"] Dec 06 04:00:00 crc kubenswrapper[4980]: I1206 04:00:00.231942 4980 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-cache-glance-default-external-api-0-cleaner-2941656ml2pt"] Dec 06 04:00:00 crc kubenswrapper[4980]: I1206 04:00:00.233137 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-cache-glance-default-external-api-0-cleaner-2941656ml2pt" Dec 06 04:00:00 crc kubenswrapper[4980]: I1206 04:00:00.244363 4980 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-cache-glance-default-external-api-0-cleaner-2941656ml2pt"] Dec 06 04:00:00 crc kubenswrapper[4980]: I1206 04:00:00.287449 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-cache-glance-default-internal-api-0-cleaner-294165644rpt\" (UID: \"659983bd-c2de-4f2f-88a4-395f494156c8\") " pod="glance-kuttl-tests/glance-cache-glance-default-internal-api-0-cleaner-294165644rpt" Dec 06 04:00:00 crc kubenswrapper[4980]: I1206 04:00:00.287558 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/df42bf97-9401-424b-b591-af4d26628458-config-volume\") pod \"collect-profiles-29416560-swv82\" (UID: \"df42bf97-9401-424b-b591-af4d26628458\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416560-swv82" Dec 06 04:00:00 crc kubenswrapper[4980]: I1206 04:00:00.288441 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l5xr7\" (UniqueName: \"kubernetes.io/projected/df42bf97-9401-424b-b591-af4d26628458-kube-api-access-l5xr7\") pod \"collect-profiles-29416560-swv82\" (UID: \"df42bf97-9401-424b-b591-af4d26628458\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416560-swv82" Dec 06 04:00:00 crc kubenswrapper[4980]: I1206 04:00:00.288490 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-cache-config-data\" (UniqueName: \"kubernetes.io/secret/659983bd-c2de-4f2f-88a4-395f494156c8-image-cache-config-data\") pod \"glance-cache-glance-default-internal-api-0-cleaner-294165644rpt\" (UID: \"659983bd-c2de-4f2f-88a4-395f494156c8\") " pod="glance-kuttl-tests/glance-cache-glance-default-internal-api-0-cleaner-294165644rpt" Dec 06 04:00:00 crc kubenswrapper[4980]: I1206 04:00:00.288515 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/df42bf97-9401-424b-b591-af4d26628458-secret-volume\") pod \"collect-profiles-29416560-swv82\" (UID: \"df42bf97-9401-424b-b591-af4d26628458\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416560-swv82" Dec 06 04:00:00 crc kubenswrapper[4980]: I1206 04:00:00.288777 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ll5vn\" (UniqueName: \"kubernetes.io/projected/659983bd-c2de-4f2f-88a4-395f494156c8-kube-api-access-ll5vn\") pod \"glance-cache-glance-default-internal-api-0-cleaner-294165644rpt\" (UID: \"659983bd-c2de-4f2f-88a4-395f494156c8\") " pod="glance-kuttl-tests/glance-cache-glance-default-internal-api-0-cleaner-294165644rpt" Dec 06 04:00:00 crc kubenswrapper[4980]: I1206 04:00:00.320142 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-cache-glance-default-internal-api-0-cleaner-294165644rpt\" (UID: \"659983bd-c2de-4f2f-88a4-395f494156c8\") " pod="glance-kuttl-tests/glance-cache-glance-default-internal-api-0-cleaner-294165644rpt" Dec 06 04:00:00 crc kubenswrapper[4980]: I1206 04:00:00.390568 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-cache-config-data\" (UniqueName: \"kubernetes.io/secret/735c2ee4-228e-498d-babc-b395ba133918-image-cache-config-data\") pod \"glance-cache-glance-default-external-api-0-cleaner-2941656ml2pt\" (UID: \"735c2ee4-228e-498d-babc-b395ba133918\") " pod="glance-kuttl-tests/glance-cache-glance-default-external-api-0-cleaner-2941656ml2pt" Dec 06 04:00:00 crc kubenswrapper[4980]: I1206 04:00:00.390687 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ll5vn\" (UniqueName: \"kubernetes.io/projected/659983bd-c2de-4f2f-88a4-395f494156c8-kube-api-access-ll5vn\") pod \"glance-cache-glance-default-internal-api-0-cleaner-294165644rpt\" (UID: \"659983bd-c2de-4f2f-88a4-395f494156c8\") " pod="glance-kuttl-tests/glance-cache-glance-default-internal-api-0-cleaner-294165644rpt" Dec 06 04:00:00 crc kubenswrapper[4980]: I1206 04:00:00.390720 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-cache-glance-default-external-api-0-cleaner-2941656ml2pt\" (UID: \"735c2ee4-228e-498d-babc-b395ba133918\") " pod="glance-kuttl-tests/glance-cache-glance-default-external-api-0-cleaner-2941656ml2pt" Dec 06 04:00:00 crc kubenswrapper[4980]: I1206 04:00:00.390763 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/df42bf97-9401-424b-b591-af4d26628458-config-volume\") pod \"collect-profiles-29416560-swv82\" (UID: \"df42bf97-9401-424b-b591-af4d26628458\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416560-swv82" Dec 06 04:00:00 crc kubenswrapper[4980]: I1206 04:00:00.390839 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l5xr7\" (UniqueName: \"kubernetes.io/projected/df42bf97-9401-424b-b591-af4d26628458-kube-api-access-l5xr7\") pod \"collect-profiles-29416560-swv82\" (UID: \"df42bf97-9401-424b-b591-af4d26628458\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416560-swv82" Dec 06 04:00:00 crc kubenswrapper[4980]: I1206 04:00:00.390866 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-cache-config-data\" (UniqueName: \"kubernetes.io/secret/659983bd-c2de-4f2f-88a4-395f494156c8-image-cache-config-data\") pod \"glance-cache-glance-default-internal-api-0-cleaner-294165644rpt\" (UID: \"659983bd-c2de-4f2f-88a4-395f494156c8\") " pod="glance-kuttl-tests/glance-cache-glance-default-internal-api-0-cleaner-294165644rpt" Dec 06 04:00:00 crc kubenswrapper[4980]: I1206 04:00:00.390893 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9mwx4\" (UniqueName: \"kubernetes.io/projected/735c2ee4-228e-498d-babc-b395ba133918-kube-api-access-9mwx4\") pod \"glance-cache-glance-default-external-api-0-cleaner-2941656ml2pt\" (UID: \"735c2ee4-228e-498d-babc-b395ba133918\") " pod="glance-kuttl-tests/glance-cache-glance-default-external-api-0-cleaner-2941656ml2pt" Dec 06 04:00:00 crc kubenswrapper[4980]: I1206 04:00:00.391166 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/df42bf97-9401-424b-b591-af4d26628458-secret-volume\") pod \"collect-profiles-29416560-swv82\" (UID: \"df42bf97-9401-424b-b591-af4d26628458\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416560-swv82" Dec 06 04:00:00 crc kubenswrapper[4980]: I1206 04:00:00.392309 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/df42bf97-9401-424b-b591-af4d26628458-config-volume\") pod \"collect-profiles-29416560-swv82\" (UID: \"df42bf97-9401-424b-b591-af4d26628458\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416560-swv82" Dec 06 04:00:00 crc kubenswrapper[4980]: I1206 04:00:00.395299 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-cache-config-data\" (UniqueName: \"kubernetes.io/secret/659983bd-c2de-4f2f-88a4-395f494156c8-image-cache-config-data\") pod \"glance-cache-glance-default-internal-api-0-cleaner-294165644rpt\" (UID: \"659983bd-c2de-4f2f-88a4-395f494156c8\") " pod="glance-kuttl-tests/glance-cache-glance-default-internal-api-0-cleaner-294165644rpt" Dec 06 04:00:00 crc kubenswrapper[4980]: I1206 04:00:00.395639 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/df42bf97-9401-424b-b591-af4d26628458-secret-volume\") pod \"collect-profiles-29416560-swv82\" (UID: \"df42bf97-9401-424b-b591-af4d26628458\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416560-swv82" Dec 06 04:00:00 crc kubenswrapper[4980]: I1206 04:00:00.408208 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ll5vn\" (UniqueName: \"kubernetes.io/projected/659983bd-c2de-4f2f-88a4-395f494156c8-kube-api-access-ll5vn\") pod \"glance-cache-glance-default-internal-api-0-cleaner-294165644rpt\" (UID: \"659983bd-c2de-4f2f-88a4-395f494156c8\") " pod="glance-kuttl-tests/glance-cache-glance-default-internal-api-0-cleaner-294165644rpt" Dec 06 04:00:00 crc kubenswrapper[4980]: I1206 04:00:00.409413 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l5xr7\" (UniqueName: \"kubernetes.io/projected/df42bf97-9401-424b-b591-af4d26628458-kube-api-access-l5xr7\") pod \"collect-profiles-29416560-swv82\" (UID: \"df42bf97-9401-424b-b591-af4d26628458\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416560-swv82" Dec 06 04:00:00 crc kubenswrapper[4980]: I1206 04:00:00.414068 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-cache-glance-default-external-api-0-cleaner-2941656ml2pt\" (UID: \"735c2ee4-228e-498d-babc-b395ba133918\") " pod="glance-kuttl-tests/glance-cache-glance-default-external-api-0-cleaner-2941656ml2pt" Dec 06 04:00:00 crc kubenswrapper[4980]: I1206 04:00:00.492951 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9mwx4\" (UniqueName: \"kubernetes.io/projected/735c2ee4-228e-498d-babc-b395ba133918-kube-api-access-9mwx4\") pod \"glance-cache-glance-default-external-api-0-cleaner-2941656ml2pt\" (UID: \"735c2ee4-228e-498d-babc-b395ba133918\") " pod="glance-kuttl-tests/glance-cache-glance-default-external-api-0-cleaner-2941656ml2pt" Dec 06 04:00:00 crc kubenswrapper[4980]: I1206 04:00:00.493025 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-cache-config-data\" (UniqueName: \"kubernetes.io/secret/735c2ee4-228e-498d-babc-b395ba133918-image-cache-config-data\") pod \"glance-cache-glance-default-external-api-0-cleaner-2941656ml2pt\" (UID: \"735c2ee4-228e-498d-babc-b395ba133918\") " pod="glance-kuttl-tests/glance-cache-glance-default-external-api-0-cleaner-2941656ml2pt" Dec 06 04:00:00 crc kubenswrapper[4980]: I1206 04:00:00.495962 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-cache-config-data\" (UniqueName: \"kubernetes.io/secret/735c2ee4-228e-498d-babc-b395ba133918-image-cache-config-data\") pod \"glance-cache-glance-default-external-api-0-cleaner-2941656ml2pt\" (UID: \"735c2ee4-228e-498d-babc-b395ba133918\") " pod="glance-kuttl-tests/glance-cache-glance-default-external-api-0-cleaner-2941656ml2pt" Dec 06 04:00:00 crc kubenswrapper[4980]: I1206 04:00:00.508079 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9mwx4\" (UniqueName: \"kubernetes.io/projected/735c2ee4-228e-498d-babc-b395ba133918-kube-api-access-9mwx4\") pod \"glance-cache-glance-default-external-api-0-cleaner-2941656ml2pt\" (UID: \"735c2ee4-228e-498d-babc-b395ba133918\") " pod="glance-kuttl-tests/glance-cache-glance-default-external-api-0-cleaner-2941656ml2pt" Dec 06 04:00:00 crc kubenswrapper[4980]: I1206 04:00:00.520095 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-cache-glance-default-internal-api-0-cleaner-294165644rpt" Dec 06 04:00:00 crc kubenswrapper[4980]: I1206 04:00:00.529996 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29416560-swv82" Dec 06 04:00:00 crc kubenswrapper[4980]: I1206 04:00:00.555147 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-cache-glance-default-external-api-0-cleaner-2941656ml2pt" Dec 06 04:00:00 crc kubenswrapper[4980]: I1206 04:00:00.952990 4980 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-cache-glance-default-internal-api-0-cleaner-294165644rpt"] Dec 06 04:00:00 crc kubenswrapper[4980]: I1206 04:00:00.990666 4980 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29416560-swv82"] Dec 06 04:00:01 crc kubenswrapper[4980]: I1206 04:00:01.036152 4980 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-cache-glance-default-external-api-0-cleaner-2941656ml2pt"] Dec 06 04:00:01 crc kubenswrapper[4980]: I1206 04:00:01.368819 4980 generic.go:334] "Generic (PLEG): container finished" podID="df42bf97-9401-424b-b591-af4d26628458" containerID="49405e851af926f72d7347f87427db4fa0b792e357cf10270dda72eb76f91226" exitCode=0 Dec 06 04:00:01 crc kubenswrapper[4980]: I1206 04:00:01.368918 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29416560-swv82" event={"ID":"df42bf97-9401-424b-b591-af4d26628458","Type":"ContainerDied","Data":"49405e851af926f72d7347f87427db4fa0b792e357cf10270dda72eb76f91226"} Dec 06 04:00:01 crc kubenswrapper[4980]: I1206 04:00:01.368957 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29416560-swv82" event={"ID":"df42bf97-9401-424b-b591-af4d26628458","Type":"ContainerStarted","Data":"270bdd793b46a2dc109721d94bb6e880c5697701600983dd7d0918dbff4781b6"} Dec 06 04:00:01 crc kubenswrapper[4980]: I1206 04:00:01.371372 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-cache-glance-default-internal-api-0-cleaner-294165644rpt" event={"ID":"659983bd-c2de-4f2f-88a4-395f494156c8","Type":"ContainerStarted","Data":"c57821ee968b32eedefe48949860e775ad2447b91121e36422a9dc792b4b0305"} Dec 06 04:00:01 crc kubenswrapper[4980]: I1206 04:00:01.373872 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-cache-glance-default-external-api-0-cleaner-2941656ml2pt" event={"ID":"735c2ee4-228e-498d-babc-b395ba133918","Type":"ContainerStarted","Data":"5a2bbcfbf4d7f3369ad037f7f0f7b971e1982966d0d0f281df5ad84c18573b86"} Dec 06 04:00:02 crc kubenswrapper[4980]: I1206 04:00:02.391366 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-cache-glance-default-external-api-0-cleaner-2941656ml2pt" event={"ID":"735c2ee4-228e-498d-babc-b395ba133918","Type":"ContainerStarted","Data":"e10584723f4619cc3c8ba8d3daabd04f544497a2f37c6cb94c31bd3f72591802"} Dec 06 04:00:02 crc kubenswrapper[4980]: I1206 04:00:02.394871 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-cache-glance-default-internal-api-0-cleaner-294165644rpt" event={"ID":"659983bd-c2de-4f2f-88a4-395f494156c8","Type":"ContainerStarted","Data":"c9d3b427cfab186f27585776a9f6f166dab9ff2a1aec42256af5bd71e99874fd"} Dec 06 04:00:02 crc kubenswrapper[4980]: I1206 04:00:02.429811 4980 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/glance-cache-glance-default-internal-api-0-cleaner-294165644rpt" podStartSLOduration=2.429782938 podStartE2EDuration="2.429782938s" podCreationTimestamp="2025-12-06 04:00:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 04:00:02.423637581 +0000 UTC m=+1561.668388892" watchObservedRunningTime="2025-12-06 04:00:02.429782938 +0000 UTC m=+1561.674534199" Dec 06 04:00:02 crc kubenswrapper[4980]: I1206 04:00:02.433621 4980 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/glance-cache-glance-default-external-api-0-cleaner-2941656ml2pt" podStartSLOduration=2.433609118 podStartE2EDuration="2.433609118s" podCreationTimestamp="2025-12-06 04:00:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 04:00:02.411014258 +0000 UTC m=+1561.655765549" watchObservedRunningTime="2025-12-06 04:00:02.433609118 +0000 UTC m=+1561.678360389" Dec 06 04:00:02 crc kubenswrapper[4980]: I1206 04:00:02.742323 4980 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29416560-swv82" Dec 06 04:00:02 crc kubenswrapper[4980]: I1206 04:00:02.823334 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/df42bf97-9401-424b-b591-af4d26628458-secret-volume\") pod \"df42bf97-9401-424b-b591-af4d26628458\" (UID: \"df42bf97-9401-424b-b591-af4d26628458\") " Dec 06 04:00:02 crc kubenswrapper[4980]: I1206 04:00:02.823720 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/df42bf97-9401-424b-b591-af4d26628458-config-volume\") pod \"df42bf97-9401-424b-b591-af4d26628458\" (UID: \"df42bf97-9401-424b-b591-af4d26628458\") " Dec 06 04:00:02 crc kubenswrapper[4980]: I1206 04:00:02.824932 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/df42bf97-9401-424b-b591-af4d26628458-config-volume" (OuterVolumeSpecName: "config-volume") pod "df42bf97-9401-424b-b591-af4d26628458" (UID: "df42bf97-9401-424b-b591-af4d26628458"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 04:00:02 crc kubenswrapper[4980]: I1206 04:00:02.825014 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l5xr7\" (UniqueName: \"kubernetes.io/projected/df42bf97-9401-424b-b591-af4d26628458-kube-api-access-l5xr7\") pod \"df42bf97-9401-424b-b591-af4d26628458\" (UID: \"df42bf97-9401-424b-b591-af4d26628458\") " Dec 06 04:00:02 crc kubenswrapper[4980]: I1206 04:00:02.825788 4980 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/df42bf97-9401-424b-b591-af4d26628458-config-volume\") on node \"crc\" DevicePath \"\"" Dec 06 04:00:02 crc kubenswrapper[4980]: I1206 04:00:02.830977 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/df42bf97-9401-424b-b591-af4d26628458-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "df42bf97-9401-424b-b591-af4d26628458" (UID: "df42bf97-9401-424b-b591-af4d26628458"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 04:00:02 crc kubenswrapper[4980]: I1206 04:00:02.831719 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/df42bf97-9401-424b-b591-af4d26628458-kube-api-access-l5xr7" (OuterVolumeSpecName: "kube-api-access-l5xr7") pod "df42bf97-9401-424b-b591-af4d26628458" (UID: "df42bf97-9401-424b-b591-af4d26628458"). InnerVolumeSpecName "kube-api-access-l5xr7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 04:00:02 crc kubenswrapper[4980]: I1206 04:00:02.927507 4980 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/df42bf97-9401-424b-b591-af4d26628458-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 06 04:00:02 crc kubenswrapper[4980]: I1206 04:00:02.927576 4980 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l5xr7\" (UniqueName: \"kubernetes.io/projected/df42bf97-9401-424b-b591-af4d26628458-kube-api-access-l5xr7\") on node \"crc\" DevicePath \"\"" Dec 06 04:00:03 crc kubenswrapper[4980]: I1206 04:00:03.406962 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29416560-swv82" event={"ID":"df42bf97-9401-424b-b591-af4d26628458","Type":"ContainerDied","Data":"270bdd793b46a2dc109721d94bb6e880c5697701600983dd7d0918dbff4781b6"} Dec 06 04:00:03 crc kubenswrapper[4980]: I1206 04:00:03.407040 4980 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="270bdd793b46a2dc109721d94bb6e880c5697701600983dd7d0918dbff4781b6" Dec 06 04:00:03 crc kubenswrapper[4980]: I1206 04:00:03.406995 4980 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29416560-swv82" Dec 06 04:00:03 crc kubenswrapper[4980]: I1206 04:00:03.411222 4980 generic.go:334] "Generic (PLEG): container finished" podID="659983bd-c2de-4f2f-88a4-395f494156c8" containerID="c9d3b427cfab186f27585776a9f6f166dab9ff2a1aec42256af5bd71e99874fd" exitCode=0 Dec 06 04:00:03 crc kubenswrapper[4980]: I1206 04:00:03.411343 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-cache-glance-default-internal-api-0-cleaner-294165644rpt" event={"ID":"659983bd-c2de-4f2f-88a4-395f494156c8","Type":"ContainerDied","Data":"c9d3b427cfab186f27585776a9f6f166dab9ff2a1aec42256af5bd71e99874fd"} Dec 06 04:00:03 crc kubenswrapper[4980]: I1206 04:00:03.413339 4980 generic.go:334] "Generic (PLEG): container finished" podID="735c2ee4-228e-498d-babc-b395ba133918" containerID="e10584723f4619cc3c8ba8d3daabd04f544497a2f37c6cb94c31bd3f72591802" exitCode=0 Dec 06 04:00:03 crc kubenswrapper[4980]: I1206 04:00:03.413370 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-cache-glance-default-external-api-0-cleaner-2941656ml2pt" event={"ID":"735c2ee4-228e-498d-babc-b395ba133918","Type":"ContainerDied","Data":"e10584723f4619cc3c8ba8d3daabd04f544497a2f37c6cb94c31bd3f72591802"} Dec 06 04:00:04 crc kubenswrapper[4980]: I1206 04:00:04.806035 4980 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-cache-glance-default-external-api-0-cleaner-2941656ml2pt" Dec 06 04:00:04 crc kubenswrapper[4980]: I1206 04:00:04.813026 4980 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-cache-glance-default-internal-api-0-cleaner-294165644rpt" Dec 06 04:00:04 crc kubenswrapper[4980]: I1206 04:00:04.857759 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-cache-config-data\" (UniqueName: \"kubernetes.io/secret/659983bd-c2de-4f2f-88a4-395f494156c8-image-cache-config-data\") pod \"659983bd-c2de-4f2f-88a4-395f494156c8\" (UID: \"659983bd-c2de-4f2f-88a4-395f494156c8\") " Dec 06 04:00:04 crc kubenswrapper[4980]: I1206 04:00:04.857828 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ll5vn\" (UniqueName: \"kubernetes.io/projected/659983bd-c2de-4f2f-88a4-395f494156c8-kube-api-access-ll5vn\") pod \"659983bd-c2de-4f2f-88a4-395f494156c8\" (UID: \"659983bd-c2de-4f2f-88a4-395f494156c8\") " Dec 06 04:00:04 crc kubenswrapper[4980]: I1206 04:00:04.857855 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance-cache\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"735c2ee4-228e-498d-babc-b395ba133918\" (UID: \"735c2ee4-228e-498d-babc-b395ba133918\") " Dec 06 04:00:04 crc kubenswrapper[4980]: I1206 04:00:04.857911 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9mwx4\" (UniqueName: \"kubernetes.io/projected/735c2ee4-228e-498d-babc-b395ba133918-kube-api-access-9mwx4\") pod \"735c2ee4-228e-498d-babc-b395ba133918\" (UID: \"735c2ee4-228e-498d-babc-b395ba133918\") " Dec 06 04:00:04 crc kubenswrapper[4980]: I1206 04:00:04.857948 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-cache-config-data\" (UniqueName: \"kubernetes.io/secret/735c2ee4-228e-498d-babc-b395ba133918-image-cache-config-data\") pod \"735c2ee4-228e-498d-babc-b395ba133918\" (UID: \"735c2ee4-228e-498d-babc-b395ba133918\") " Dec 06 04:00:04 crc kubenswrapper[4980]: I1206 04:00:04.858026 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance-cache\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"659983bd-c2de-4f2f-88a4-395f494156c8\" (UID: \"659983bd-c2de-4f2f-88a4-395f494156c8\") " Dec 06 04:00:04 crc kubenswrapper[4980]: I1206 04:00:04.864501 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage01-crc" (OuterVolumeSpecName: "glance-cache") pod "659983bd-c2de-4f2f-88a4-395f494156c8" (UID: "659983bd-c2de-4f2f-88a4-395f494156c8"). InnerVolumeSpecName "local-storage01-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 06 04:00:04 crc kubenswrapper[4980]: I1206 04:00:04.864514 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/659983bd-c2de-4f2f-88a4-395f494156c8-image-cache-config-data" (OuterVolumeSpecName: "image-cache-config-data") pod "659983bd-c2de-4f2f-88a4-395f494156c8" (UID: "659983bd-c2de-4f2f-88a4-395f494156c8"). InnerVolumeSpecName "image-cache-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 04:00:04 crc kubenswrapper[4980]: I1206 04:00:04.864897 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage10-crc" (OuterVolumeSpecName: "glance-cache") pod "735c2ee4-228e-498d-babc-b395ba133918" (UID: "735c2ee4-228e-498d-babc-b395ba133918"). InnerVolumeSpecName "local-storage10-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 06 04:00:04 crc kubenswrapper[4980]: I1206 04:00:04.865377 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/735c2ee4-228e-498d-babc-b395ba133918-kube-api-access-9mwx4" (OuterVolumeSpecName: "kube-api-access-9mwx4") pod "735c2ee4-228e-498d-babc-b395ba133918" (UID: "735c2ee4-228e-498d-babc-b395ba133918"). InnerVolumeSpecName "kube-api-access-9mwx4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 04:00:04 crc kubenswrapper[4980]: I1206 04:00:04.867477 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/735c2ee4-228e-498d-babc-b395ba133918-image-cache-config-data" (OuterVolumeSpecName: "image-cache-config-data") pod "735c2ee4-228e-498d-babc-b395ba133918" (UID: "735c2ee4-228e-498d-babc-b395ba133918"). InnerVolumeSpecName "image-cache-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 04:00:04 crc kubenswrapper[4980]: I1206 04:00:04.868889 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/659983bd-c2de-4f2f-88a4-395f494156c8-kube-api-access-ll5vn" (OuterVolumeSpecName: "kube-api-access-ll5vn") pod "659983bd-c2de-4f2f-88a4-395f494156c8" (UID: "659983bd-c2de-4f2f-88a4-395f494156c8"). InnerVolumeSpecName "kube-api-access-ll5vn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 04:00:04 crc kubenswrapper[4980]: I1206 04:00:04.959400 4980 reconciler_common.go:293] "Volume detached for volume \"image-cache-config-data\" (UniqueName: \"kubernetes.io/secret/659983bd-c2de-4f2f-88a4-395f494156c8-image-cache-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 04:00:04 crc kubenswrapper[4980]: I1206 04:00:04.959453 4980 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ll5vn\" (UniqueName: \"kubernetes.io/projected/659983bd-c2de-4f2f-88a4-395f494156c8-kube-api-access-ll5vn\") on node \"crc\" DevicePath \"\"" Dec 06 04:00:04 crc kubenswrapper[4980]: I1206 04:00:04.959465 4980 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9mwx4\" (UniqueName: \"kubernetes.io/projected/735c2ee4-228e-498d-babc-b395ba133918-kube-api-access-9mwx4\") on node \"crc\" DevicePath \"\"" Dec 06 04:00:04 crc kubenswrapper[4980]: I1206 04:00:04.959473 4980 reconciler_common.go:293] "Volume detached for volume \"image-cache-config-data\" (UniqueName: \"kubernetes.io/secret/735c2ee4-228e-498d-babc-b395ba133918-image-cache-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 04:00:05 crc kubenswrapper[4980]: I1206 04:00:05.432646 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-cache-glance-default-internal-api-0-cleaner-294165644rpt" event={"ID":"659983bd-c2de-4f2f-88a4-395f494156c8","Type":"ContainerDied","Data":"c57821ee968b32eedefe48949860e775ad2447b91121e36422a9dc792b4b0305"} Dec 06 04:00:05 crc kubenswrapper[4980]: I1206 04:00:05.432754 4980 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c57821ee968b32eedefe48949860e775ad2447b91121e36422a9dc792b4b0305" Dec 06 04:00:05 crc kubenswrapper[4980]: I1206 04:00:05.432684 4980 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-cache-glance-default-internal-api-0-cleaner-294165644rpt" Dec 06 04:00:05 crc kubenswrapper[4980]: I1206 04:00:05.434879 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-cache-glance-default-external-api-0-cleaner-2941656ml2pt" event={"ID":"735c2ee4-228e-498d-babc-b395ba133918","Type":"ContainerDied","Data":"5a2bbcfbf4d7f3369ad037f7f0f7b971e1982966d0d0f281df5ad84c18573b86"} Dec 06 04:00:05 crc kubenswrapper[4980]: I1206 04:00:05.434927 4980 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5a2bbcfbf4d7f3369ad037f7f0f7b971e1982966d0d0f281df5ad84c18573b86" Dec 06 04:00:05 crc kubenswrapper[4980]: I1206 04:00:05.434975 4980 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-cache-glance-default-external-api-0-cleaner-2941656ml2pt" Dec 06 04:00:08 crc kubenswrapper[4980]: I1206 04:00:08.184953 4980 scope.go:117] "RemoveContainer" containerID="e8878e13c398cceec46bb52c4c9ea0e06dbf4317e6ad39b69ca763a0d8b15497" Dec 06 04:00:08 crc kubenswrapper[4980]: E1206 04:00:08.185677 4980 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-r5zfb_openshift-machine-config-operator(320f44d1-a671-4a91-b328-a8b0fdd8f23a)\"" pod="openshift-machine-config-operator/machine-config-daemon-r5zfb" podUID="320f44d1-a671-4a91-b328-a8b0fdd8f23a" Dec 06 04:00:21 crc kubenswrapper[4980]: I1206 04:00:21.188666 4980 scope.go:117] "RemoveContainer" containerID="e8878e13c398cceec46bb52c4c9ea0e06dbf4317e6ad39b69ca763a0d8b15497" Dec 06 04:00:21 crc kubenswrapper[4980]: E1206 04:00:21.189224 4980 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-r5zfb_openshift-machine-config-operator(320f44d1-a671-4a91-b328-a8b0fdd8f23a)\"" pod="openshift-machine-config-operator/machine-config-daemon-r5zfb" podUID="320f44d1-a671-4a91-b328-a8b0fdd8f23a" Dec 06 04:00:35 crc kubenswrapper[4980]: I1206 04:00:35.185364 4980 scope.go:117] "RemoveContainer" containerID="e8878e13c398cceec46bb52c4c9ea0e06dbf4317e6ad39b69ca763a0d8b15497" Dec 06 04:00:35 crc kubenswrapper[4980]: E1206 04:00:35.186186 4980 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-r5zfb_openshift-machine-config-operator(320f44d1-a671-4a91-b328-a8b0fdd8f23a)\"" pod="openshift-machine-config-operator/machine-config-daemon-r5zfb" podUID="320f44d1-a671-4a91-b328-a8b0fdd8f23a" Dec 06 04:00:47 crc kubenswrapper[4980]: I1206 04:00:47.184734 4980 scope.go:117] "RemoveContainer" containerID="e8878e13c398cceec46bb52c4c9ea0e06dbf4317e6ad39b69ca763a0d8b15497" Dec 06 04:00:47 crc kubenswrapper[4980]: E1206 04:00:47.185551 4980 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-r5zfb_openshift-machine-config-operator(320f44d1-a671-4a91-b328-a8b0fdd8f23a)\"" pod="openshift-machine-config-operator/machine-config-daemon-r5zfb" podUID="320f44d1-a671-4a91-b328-a8b0fdd8f23a" Dec 06 04:01:00 crc kubenswrapper[4980]: I1206 04:01:00.153743 4980 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/keystone-cron-29416561-bcltb"] Dec 06 04:01:00 crc kubenswrapper[4980]: E1206 04:01:00.156426 4980 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="659983bd-c2de-4f2f-88a4-395f494156c8" containerName="glance-cache-glance-default-internal-api-0-cleaner" Dec 06 04:01:00 crc kubenswrapper[4980]: I1206 04:01:00.156567 4980 state_mem.go:107] "Deleted CPUSet assignment" podUID="659983bd-c2de-4f2f-88a4-395f494156c8" containerName="glance-cache-glance-default-internal-api-0-cleaner" Dec 06 04:01:00 crc kubenswrapper[4980]: E1206 04:01:00.156712 4980 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="df42bf97-9401-424b-b591-af4d26628458" containerName="collect-profiles" Dec 06 04:01:00 crc kubenswrapper[4980]: I1206 04:01:00.156794 4980 state_mem.go:107] "Deleted CPUSet assignment" podUID="df42bf97-9401-424b-b591-af4d26628458" containerName="collect-profiles" Dec 06 04:01:00 crc kubenswrapper[4980]: E1206 04:01:00.156888 4980 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="735c2ee4-228e-498d-babc-b395ba133918" containerName="glance-cache-glance-default-external-api-0-cleaner" Dec 06 04:01:00 crc kubenswrapper[4980]: I1206 04:01:00.156963 4980 state_mem.go:107] "Deleted CPUSet assignment" podUID="735c2ee4-228e-498d-babc-b395ba133918" containerName="glance-cache-glance-default-external-api-0-cleaner" Dec 06 04:01:00 crc kubenswrapper[4980]: I1206 04:01:00.157330 4980 memory_manager.go:354] "RemoveStaleState removing state" podUID="735c2ee4-228e-498d-babc-b395ba133918" containerName="glance-cache-glance-default-external-api-0-cleaner" Dec 06 04:01:00 crc kubenswrapper[4980]: I1206 04:01:00.157435 4980 memory_manager.go:354] "RemoveStaleState removing state" podUID="659983bd-c2de-4f2f-88a4-395f494156c8" containerName="glance-cache-glance-default-internal-api-0-cleaner" Dec 06 04:01:00 crc kubenswrapper[4980]: I1206 04:01:00.157589 4980 memory_manager.go:354] "RemoveStaleState removing state" podUID="df42bf97-9401-424b-b591-af4d26628458" containerName="collect-profiles" Dec 06 04:01:00 crc kubenswrapper[4980]: I1206 04:01:00.158328 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/keystone-cron-29416561-bcltb" Dec 06 04:01:00 crc kubenswrapper[4980]: I1206 04:01:00.171298 4980 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/keystone-cron-29416561-bcltb"] Dec 06 04:01:00 crc kubenswrapper[4980]: I1206 04:01:00.191776 4980 scope.go:117] "RemoveContainer" containerID="e8878e13c398cceec46bb52c4c9ea0e06dbf4317e6ad39b69ca763a0d8b15497" Dec 06 04:01:00 crc kubenswrapper[4980]: E1206 04:01:00.191998 4980 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-r5zfb_openshift-machine-config-operator(320f44d1-a671-4a91-b328-a8b0fdd8f23a)\"" pod="openshift-machine-config-operator/machine-config-daemon-r5zfb" podUID="320f44d1-a671-4a91-b328-a8b0fdd8f23a" Dec 06 04:01:00 crc kubenswrapper[4980]: I1206 04:01:00.292340 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/786dd383-8c53-46e3-afc3-eebbf6ecae33-config-data\") pod \"keystone-cron-29416561-bcltb\" (UID: \"786dd383-8c53-46e3-afc3-eebbf6ecae33\") " pod="glance-kuttl-tests/keystone-cron-29416561-bcltb" Dec 06 04:01:00 crc kubenswrapper[4980]: I1206 04:01:00.292556 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/786dd383-8c53-46e3-afc3-eebbf6ecae33-fernet-keys\") pod \"keystone-cron-29416561-bcltb\" (UID: \"786dd383-8c53-46e3-afc3-eebbf6ecae33\") " pod="glance-kuttl-tests/keystone-cron-29416561-bcltb" Dec 06 04:01:00 crc kubenswrapper[4980]: I1206 04:01:00.292796 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kc2zt\" (UniqueName: \"kubernetes.io/projected/786dd383-8c53-46e3-afc3-eebbf6ecae33-kube-api-access-kc2zt\") pod \"keystone-cron-29416561-bcltb\" (UID: \"786dd383-8c53-46e3-afc3-eebbf6ecae33\") " pod="glance-kuttl-tests/keystone-cron-29416561-bcltb" Dec 06 04:01:00 crc kubenswrapper[4980]: I1206 04:01:00.394954 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kc2zt\" (UniqueName: \"kubernetes.io/projected/786dd383-8c53-46e3-afc3-eebbf6ecae33-kube-api-access-kc2zt\") pod \"keystone-cron-29416561-bcltb\" (UID: \"786dd383-8c53-46e3-afc3-eebbf6ecae33\") " pod="glance-kuttl-tests/keystone-cron-29416561-bcltb" Dec 06 04:01:00 crc kubenswrapper[4980]: I1206 04:01:00.395081 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/786dd383-8c53-46e3-afc3-eebbf6ecae33-config-data\") pod \"keystone-cron-29416561-bcltb\" (UID: \"786dd383-8c53-46e3-afc3-eebbf6ecae33\") " pod="glance-kuttl-tests/keystone-cron-29416561-bcltb" Dec 06 04:01:00 crc kubenswrapper[4980]: I1206 04:01:00.395159 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/786dd383-8c53-46e3-afc3-eebbf6ecae33-fernet-keys\") pod \"keystone-cron-29416561-bcltb\" (UID: \"786dd383-8c53-46e3-afc3-eebbf6ecae33\") " pod="glance-kuttl-tests/keystone-cron-29416561-bcltb" Dec 06 04:01:00 crc kubenswrapper[4980]: I1206 04:01:00.404464 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/786dd383-8c53-46e3-afc3-eebbf6ecae33-fernet-keys\") pod \"keystone-cron-29416561-bcltb\" (UID: \"786dd383-8c53-46e3-afc3-eebbf6ecae33\") " pod="glance-kuttl-tests/keystone-cron-29416561-bcltb" Dec 06 04:01:00 crc kubenswrapper[4980]: I1206 04:01:00.404604 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/786dd383-8c53-46e3-afc3-eebbf6ecae33-config-data\") pod \"keystone-cron-29416561-bcltb\" (UID: \"786dd383-8c53-46e3-afc3-eebbf6ecae33\") " pod="glance-kuttl-tests/keystone-cron-29416561-bcltb" Dec 06 04:01:00 crc kubenswrapper[4980]: I1206 04:01:00.436434 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kc2zt\" (UniqueName: \"kubernetes.io/projected/786dd383-8c53-46e3-afc3-eebbf6ecae33-kube-api-access-kc2zt\") pod \"keystone-cron-29416561-bcltb\" (UID: \"786dd383-8c53-46e3-afc3-eebbf6ecae33\") " pod="glance-kuttl-tests/keystone-cron-29416561-bcltb" Dec 06 04:01:00 crc kubenswrapper[4980]: I1206 04:01:00.499058 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/keystone-cron-29416561-bcltb" Dec 06 04:01:01 crc kubenswrapper[4980]: I1206 04:01:01.000018 4980 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/keystone-cron-29416561-bcltb"] Dec 06 04:01:01 crc kubenswrapper[4980]: I1206 04:01:01.120584 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/keystone-cron-29416561-bcltb" event={"ID":"786dd383-8c53-46e3-afc3-eebbf6ecae33","Type":"ContainerStarted","Data":"b9b853c4f4f574a5d4df2eb0a5a2329944466c8ef3edea05fd9b8cd90c668629"} Dec 06 04:01:02 crc kubenswrapper[4980]: I1206 04:01:02.132274 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/keystone-cron-29416561-bcltb" event={"ID":"786dd383-8c53-46e3-afc3-eebbf6ecae33","Type":"ContainerStarted","Data":"59124c6f43ec1aa6d4011225b74a3df60a950e73c43b6de15409a222bb2d0658"} Dec 06 04:01:02 crc kubenswrapper[4980]: I1206 04:01:02.152378 4980 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/keystone-cron-29416561-bcltb" podStartSLOduration=2.152332785 podStartE2EDuration="2.152332785s" podCreationTimestamp="2025-12-06 04:01:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 04:01:02.149974748 +0000 UTC m=+1621.394726029" watchObservedRunningTime="2025-12-06 04:01:02.152332785 +0000 UTC m=+1621.397084056" Dec 06 04:01:04 crc kubenswrapper[4980]: I1206 04:01:04.149771 4980 generic.go:334] "Generic (PLEG): container finished" podID="786dd383-8c53-46e3-afc3-eebbf6ecae33" containerID="59124c6f43ec1aa6d4011225b74a3df60a950e73c43b6de15409a222bb2d0658" exitCode=0 Dec 06 04:01:04 crc kubenswrapper[4980]: I1206 04:01:04.149884 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/keystone-cron-29416561-bcltb" event={"ID":"786dd383-8c53-46e3-afc3-eebbf6ecae33","Type":"ContainerDied","Data":"59124c6f43ec1aa6d4011225b74a3df60a950e73c43b6de15409a222bb2d0658"} Dec 06 04:01:05 crc kubenswrapper[4980]: I1206 04:01:05.507871 4980 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/keystone-cron-29416561-bcltb" Dec 06 04:01:05 crc kubenswrapper[4980]: I1206 04:01:05.695996 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/786dd383-8c53-46e3-afc3-eebbf6ecae33-config-data\") pod \"786dd383-8c53-46e3-afc3-eebbf6ecae33\" (UID: \"786dd383-8c53-46e3-afc3-eebbf6ecae33\") " Dec 06 04:01:05 crc kubenswrapper[4980]: I1206 04:01:05.696047 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/786dd383-8c53-46e3-afc3-eebbf6ecae33-fernet-keys\") pod \"786dd383-8c53-46e3-afc3-eebbf6ecae33\" (UID: \"786dd383-8c53-46e3-afc3-eebbf6ecae33\") " Dec 06 04:01:05 crc kubenswrapper[4980]: I1206 04:01:05.696088 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kc2zt\" (UniqueName: \"kubernetes.io/projected/786dd383-8c53-46e3-afc3-eebbf6ecae33-kube-api-access-kc2zt\") pod \"786dd383-8c53-46e3-afc3-eebbf6ecae33\" (UID: \"786dd383-8c53-46e3-afc3-eebbf6ecae33\") " Dec 06 04:01:05 crc kubenswrapper[4980]: I1206 04:01:05.700762 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/786dd383-8c53-46e3-afc3-eebbf6ecae33-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "786dd383-8c53-46e3-afc3-eebbf6ecae33" (UID: "786dd383-8c53-46e3-afc3-eebbf6ecae33"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 04:01:05 crc kubenswrapper[4980]: I1206 04:01:05.700914 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/786dd383-8c53-46e3-afc3-eebbf6ecae33-kube-api-access-kc2zt" (OuterVolumeSpecName: "kube-api-access-kc2zt") pod "786dd383-8c53-46e3-afc3-eebbf6ecae33" (UID: "786dd383-8c53-46e3-afc3-eebbf6ecae33"). InnerVolumeSpecName "kube-api-access-kc2zt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 04:01:05 crc kubenswrapper[4980]: I1206 04:01:05.743851 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/786dd383-8c53-46e3-afc3-eebbf6ecae33-config-data" (OuterVolumeSpecName: "config-data") pod "786dd383-8c53-46e3-afc3-eebbf6ecae33" (UID: "786dd383-8c53-46e3-afc3-eebbf6ecae33"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 04:01:05 crc kubenswrapper[4980]: I1206 04:01:05.798040 4980 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/786dd383-8c53-46e3-afc3-eebbf6ecae33-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 04:01:05 crc kubenswrapper[4980]: I1206 04:01:05.798424 4980 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/786dd383-8c53-46e3-afc3-eebbf6ecae33-fernet-keys\") on node \"crc\" DevicePath \"\"" Dec 06 04:01:05 crc kubenswrapper[4980]: I1206 04:01:05.798435 4980 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kc2zt\" (UniqueName: \"kubernetes.io/projected/786dd383-8c53-46e3-afc3-eebbf6ecae33-kube-api-access-kc2zt\") on node \"crc\" DevicePath \"\"" Dec 06 04:01:06 crc kubenswrapper[4980]: I1206 04:01:06.172294 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/keystone-cron-29416561-bcltb" event={"ID":"786dd383-8c53-46e3-afc3-eebbf6ecae33","Type":"ContainerDied","Data":"b9b853c4f4f574a5d4df2eb0a5a2329944466c8ef3edea05fd9b8cd90c668629"} Dec 06 04:01:06 crc kubenswrapper[4980]: I1206 04:01:06.172360 4980 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b9b853c4f4f574a5d4df2eb0a5a2329944466c8ef3edea05fd9b8cd90c668629" Dec 06 04:01:06 crc kubenswrapper[4980]: I1206 04:01:06.172426 4980 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/keystone-cron-29416561-bcltb" Dec 06 04:01:06 crc kubenswrapper[4980]: I1206 04:01:06.329358 4980 scope.go:117] "RemoveContainer" containerID="e372798becf1c32343fcbf6a7cf618cccedec38ee0dce588b1a6f1db47fc3d6e" Dec 06 04:01:06 crc kubenswrapper[4980]: I1206 04:01:06.351601 4980 scope.go:117] "RemoveContainer" containerID="4e6f6016172688199f5ee8bb2fbe6e702d804a3bf782f709b474dbfe051a3c94" Dec 06 04:01:13 crc kubenswrapper[4980]: I1206 04:01:13.184831 4980 scope.go:117] "RemoveContainer" containerID="e8878e13c398cceec46bb52c4c9ea0e06dbf4317e6ad39b69ca763a0d8b15497" Dec 06 04:01:13 crc kubenswrapper[4980]: E1206 04:01:13.185817 4980 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-r5zfb_openshift-machine-config-operator(320f44d1-a671-4a91-b328-a8b0fdd8f23a)\"" pod="openshift-machine-config-operator/machine-config-daemon-r5zfb" podUID="320f44d1-a671-4a91-b328-a8b0fdd8f23a" Dec 06 04:01:28 crc kubenswrapper[4980]: I1206 04:01:28.185333 4980 scope.go:117] "RemoveContainer" containerID="e8878e13c398cceec46bb52c4c9ea0e06dbf4317e6ad39b69ca763a0d8b15497" Dec 06 04:01:28 crc kubenswrapper[4980]: E1206 04:01:28.186126 4980 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-r5zfb_openshift-machine-config-operator(320f44d1-a671-4a91-b328-a8b0fdd8f23a)\"" pod="openshift-machine-config-operator/machine-config-daemon-r5zfb" podUID="320f44d1-a671-4a91-b328-a8b0fdd8f23a" Dec 06 04:01:29 crc kubenswrapper[4980]: I1206 04:01:29.638534 4980 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-6fqt7"] Dec 06 04:01:29 crc kubenswrapper[4980]: E1206 04:01:29.638907 4980 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="786dd383-8c53-46e3-afc3-eebbf6ecae33" containerName="keystone-cron" Dec 06 04:01:29 crc kubenswrapper[4980]: I1206 04:01:29.638919 4980 state_mem.go:107] "Deleted CPUSet assignment" podUID="786dd383-8c53-46e3-afc3-eebbf6ecae33" containerName="keystone-cron" Dec 06 04:01:29 crc kubenswrapper[4980]: I1206 04:01:29.639090 4980 memory_manager.go:354] "RemoveStaleState removing state" podUID="786dd383-8c53-46e3-afc3-eebbf6ecae33" containerName="keystone-cron" Dec 06 04:01:29 crc kubenswrapper[4980]: I1206 04:01:29.640133 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-6fqt7" Dec 06 04:01:29 crc kubenswrapper[4980]: I1206 04:01:29.648126 4980 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-6fqt7"] Dec 06 04:01:29 crc kubenswrapper[4980]: I1206 04:01:29.667337 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9e5ca126-1dff-4106-8643-20657251eeea-utilities\") pod \"community-operators-6fqt7\" (UID: \"9e5ca126-1dff-4106-8643-20657251eeea\") " pod="openshift-marketplace/community-operators-6fqt7" Dec 06 04:01:29 crc kubenswrapper[4980]: I1206 04:01:29.667387 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b9qkn\" (UniqueName: \"kubernetes.io/projected/9e5ca126-1dff-4106-8643-20657251eeea-kube-api-access-b9qkn\") pod \"community-operators-6fqt7\" (UID: \"9e5ca126-1dff-4106-8643-20657251eeea\") " pod="openshift-marketplace/community-operators-6fqt7" Dec 06 04:01:29 crc kubenswrapper[4980]: I1206 04:01:29.667482 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9e5ca126-1dff-4106-8643-20657251eeea-catalog-content\") pod \"community-operators-6fqt7\" (UID: \"9e5ca126-1dff-4106-8643-20657251eeea\") " pod="openshift-marketplace/community-operators-6fqt7" Dec 06 04:01:29 crc kubenswrapper[4980]: I1206 04:01:29.769289 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9e5ca126-1dff-4106-8643-20657251eeea-catalog-content\") pod \"community-operators-6fqt7\" (UID: \"9e5ca126-1dff-4106-8643-20657251eeea\") " pod="openshift-marketplace/community-operators-6fqt7" Dec 06 04:01:29 crc kubenswrapper[4980]: I1206 04:01:29.769420 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9e5ca126-1dff-4106-8643-20657251eeea-utilities\") pod \"community-operators-6fqt7\" (UID: \"9e5ca126-1dff-4106-8643-20657251eeea\") " pod="openshift-marketplace/community-operators-6fqt7" Dec 06 04:01:29 crc kubenswrapper[4980]: I1206 04:01:29.769452 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b9qkn\" (UniqueName: \"kubernetes.io/projected/9e5ca126-1dff-4106-8643-20657251eeea-kube-api-access-b9qkn\") pod \"community-operators-6fqt7\" (UID: \"9e5ca126-1dff-4106-8643-20657251eeea\") " pod="openshift-marketplace/community-operators-6fqt7" Dec 06 04:01:29 crc kubenswrapper[4980]: I1206 04:01:29.770382 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9e5ca126-1dff-4106-8643-20657251eeea-catalog-content\") pod \"community-operators-6fqt7\" (UID: \"9e5ca126-1dff-4106-8643-20657251eeea\") " pod="openshift-marketplace/community-operators-6fqt7" Dec 06 04:01:29 crc kubenswrapper[4980]: I1206 04:01:29.771592 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9e5ca126-1dff-4106-8643-20657251eeea-utilities\") pod \"community-operators-6fqt7\" (UID: \"9e5ca126-1dff-4106-8643-20657251eeea\") " pod="openshift-marketplace/community-operators-6fqt7" Dec 06 04:01:29 crc kubenswrapper[4980]: I1206 04:01:29.797472 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b9qkn\" (UniqueName: \"kubernetes.io/projected/9e5ca126-1dff-4106-8643-20657251eeea-kube-api-access-b9qkn\") pod \"community-operators-6fqt7\" (UID: \"9e5ca126-1dff-4106-8643-20657251eeea\") " pod="openshift-marketplace/community-operators-6fqt7" Dec 06 04:01:29 crc kubenswrapper[4980]: I1206 04:01:29.964471 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-6fqt7" Dec 06 04:01:30 crc kubenswrapper[4980]: I1206 04:01:30.519167 4980 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-6fqt7"] Dec 06 04:01:31 crc kubenswrapper[4980]: I1206 04:01:31.401760 4980 generic.go:334] "Generic (PLEG): container finished" podID="9e5ca126-1dff-4106-8643-20657251eeea" containerID="3b073f68b1237680b6011a2ec68bb210106ab2eab774d382b7265eec49ca387d" exitCode=0 Dec 06 04:01:31 crc kubenswrapper[4980]: I1206 04:01:31.401830 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6fqt7" event={"ID":"9e5ca126-1dff-4106-8643-20657251eeea","Type":"ContainerDied","Data":"3b073f68b1237680b6011a2ec68bb210106ab2eab774d382b7265eec49ca387d"} Dec 06 04:01:31 crc kubenswrapper[4980]: I1206 04:01:31.401893 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6fqt7" event={"ID":"9e5ca126-1dff-4106-8643-20657251eeea","Type":"ContainerStarted","Data":"85e3e0743b6c56700dbdcc833dde3869fb6814e72a123629be7701df90616ff6"} Dec 06 04:01:32 crc kubenswrapper[4980]: I1206 04:01:32.413167 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6fqt7" event={"ID":"9e5ca126-1dff-4106-8643-20657251eeea","Type":"ContainerStarted","Data":"52f7c348bfb7fbab86d8644f90533f55ed6f68d5ffe399cb6e89ad6d81066d08"} Dec 06 04:01:33 crc kubenswrapper[4980]: I1206 04:01:33.422631 4980 generic.go:334] "Generic (PLEG): container finished" podID="9e5ca126-1dff-4106-8643-20657251eeea" containerID="52f7c348bfb7fbab86d8644f90533f55ed6f68d5ffe399cb6e89ad6d81066d08" exitCode=0 Dec 06 04:01:33 crc kubenswrapper[4980]: I1206 04:01:33.422781 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6fqt7" event={"ID":"9e5ca126-1dff-4106-8643-20657251eeea","Type":"ContainerDied","Data":"52f7c348bfb7fbab86d8644f90533f55ed6f68d5ffe399cb6e89ad6d81066d08"} Dec 06 04:01:34 crc kubenswrapper[4980]: I1206 04:01:34.433769 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6fqt7" event={"ID":"9e5ca126-1dff-4106-8643-20657251eeea","Type":"ContainerStarted","Data":"80aedde12ae1c998bed1310ca202dd66299e75b51687bcd0b4dba2c3e8fb1aba"} Dec 06 04:01:34 crc kubenswrapper[4980]: I1206 04:01:34.470973 4980 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-6fqt7" podStartSLOduration=2.942081323 podStartE2EDuration="5.470947047s" podCreationTimestamp="2025-12-06 04:01:29 +0000 UTC" firstStartedPulling="2025-12-06 04:01:31.405491702 +0000 UTC m=+1650.650243023" lastFinishedPulling="2025-12-06 04:01:33.934357436 +0000 UTC m=+1653.179108747" observedRunningTime="2025-12-06 04:01:34.456844502 +0000 UTC m=+1653.701595853" watchObservedRunningTime="2025-12-06 04:01:34.470947047 +0000 UTC m=+1653.715698358" Dec 06 04:01:39 crc kubenswrapper[4980]: I1206 04:01:39.185482 4980 scope.go:117] "RemoveContainer" containerID="e8878e13c398cceec46bb52c4c9ea0e06dbf4317e6ad39b69ca763a0d8b15497" Dec 06 04:01:39 crc kubenswrapper[4980]: E1206 04:01:39.186716 4980 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-r5zfb_openshift-machine-config-operator(320f44d1-a671-4a91-b328-a8b0fdd8f23a)\"" pod="openshift-machine-config-operator/machine-config-daemon-r5zfb" podUID="320f44d1-a671-4a91-b328-a8b0fdd8f23a" Dec 06 04:01:39 crc kubenswrapper[4980]: I1206 04:01:39.988797 4980 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-6fqt7" Dec 06 04:01:39 crc kubenswrapper[4980]: I1206 04:01:39.989696 4980 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-6fqt7" Dec 06 04:01:40 crc kubenswrapper[4980]: I1206 04:01:40.043948 4980 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-6fqt7" Dec 06 04:01:40 crc kubenswrapper[4980]: I1206 04:01:40.574331 4980 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-6fqt7" Dec 06 04:01:40 crc kubenswrapper[4980]: I1206 04:01:40.651670 4980 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-6fqt7"] Dec 06 04:01:42 crc kubenswrapper[4980]: I1206 04:01:42.513944 4980 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-6fqt7" podUID="9e5ca126-1dff-4106-8643-20657251eeea" containerName="registry-server" containerID="cri-o://80aedde12ae1c998bed1310ca202dd66299e75b51687bcd0b4dba2c3e8fb1aba" gracePeriod=2 Dec 06 04:01:42 crc kubenswrapper[4980]: I1206 04:01:42.994685 4980 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-6fqt7" Dec 06 04:01:43 crc kubenswrapper[4980]: I1206 04:01:43.095855 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9e5ca126-1dff-4106-8643-20657251eeea-utilities\") pod \"9e5ca126-1dff-4106-8643-20657251eeea\" (UID: \"9e5ca126-1dff-4106-8643-20657251eeea\") " Dec 06 04:01:43 crc kubenswrapper[4980]: I1206 04:01:43.095945 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9e5ca126-1dff-4106-8643-20657251eeea-catalog-content\") pod \"9e5ca126-1dff-4106-8643-20657251eeea\" (UID: \"9e5ca126-1dff-4106-8643-20657251eeea\") " Dec 06 04:01:43 crc kubenswrapper[4980]: I1206 04:01:43.095971 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b9qkn\" (UniqueName: \"kubernetes.io/projected/9e5ca126-1dff-4106-8643-20657251eeea-kube-api-access-b9qkn\") pod \"9e5ca126-1dff-4106-8643-20657251eeea\" (UID: \"9e5ca126-1dff-4106-8643-20657251eeea\") " Dec 06 04:01:43 crc kubenswrapper[4980]: I1206 04:01:43.097784 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9e5ca126-1dff-4106-8643-20657251eeea-utilities" (OuterVolumeSpecName: "utilities") pod "9e5ca126-1dff-4106-8643-20657251eeea" (UID: "9e5ca126-1dff-4106-8643-20657251eeea"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 04:01:43 crc kubenswrapper[4980]: I1206 04:01:43.102027 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9e5ca126-1dff-4106-8643-20657251eeea-kube-api-access-b9qkn" (OuterVolumeSpecName: "kube-api-access-b9qkn") pod "9e5ca126-1dff-4106-8643-20657251eeea" (UID: "9e5ca126-1dff-4106-8643-20657251eeea"). InnerVolumeSpecName "kube-api-access-b9qkn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 04:01:43 crc kubenswrapper[4980]: I1206 04:01:43.151613 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9e5ca126-1dff-4106-8643-20657251eeea-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "9e5ca126-1dff-4106-8643-20657251eeea" (UID: "9e5ca126-1dff-4106-8643-20657251eeea"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 04:01:43 crc kubenswrapper[4980]: I1206 04:01:43.197405 4980 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9e5ca126-1dff-4106-8643-20657251eeea-utilities\") on node \"crc\" DevicePath \"\"" Dec 06 04:01:43 crc kubenswrapper[4980]: I1206 04:01:43.197443 4980 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9e5ca126-1dff-4106-8643-20657251eeea-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 06 04:01:43 crc kubenswrapper[4980]: I1206 04:01:43.197464 4980 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b9qkn\" (UniqueName: \"kubernetes.io/projected/9e5ca126-1dff-4106-8643-20657251eeea-kube-api-access-b9qkn\") on node \"crc\" DevicePath \"\"" Dec 06 04:01:43 crc kubenswrapper[4980]: I1206 04:01:43.529157 4980 generic.go:334] "Generic (PLEG): container finished" podID="9e5ca126-1dff-4106-8643-20657251eeea" containerID="80aedde12ae1c998bed1310ca202dd66299e75b51687bcd0b4dba2c3e8fb1aba" exitCode=0 Dec 06 04:01:43 crc kubenswrapper[4980]: I1206 04:01:43.529226 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6fqt7" event={"ID":"9e5ca126-1dff-4106-8643-20657251eeea","Type":"ContainerDied","Data":"80aedde12ae1c998bed1310ca202dd66299e75b51687bcd0b4dba2c3e8fb1aba"} Dec 06 04:01:43 crc kubenswrapper[4980]: I1206 04:01:43.529268 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6fqt7" event={"ID":"9e5ca126-1dff-4106-8643-20657251eeea","Type":"ContainerDied","Data":"85e3e0743b6c56700dbdcc833dde3869fb6814e72a123629be7701df90616ff6"} Dec 06 04:01:43 crc kubenswrapper[4980]: I1206 04:01:43.529288 4980 scope.go:117] "RemoveContainer" containerID="80aedde12ae1c998bed1310ca202dd66299e75b51687bcd0b4dba2c3e8fb1aba" Dec 06 04:01:43 crc kubenswrapper[4980]: I1206 04:01:43.529354 4980 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-6fqt7" Dec 06 04:01:43 crc kubenswrapper[4980]: I1206 04:01:43.558606 4980 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-6fqt7"] Dec 06 04:01:43 crc kubenswrapper[4980]: I1206 04:01:43.568407 4980 scope.go:117] "RemoveContainer" containerID="52f7c348bfb7fbab86d8644f90533f55ed6f68d5ffe399cb6e89ad6d81066d08" Dec 06 04:01:43 crc kubenswrapper[4980]: I1206 04:01:43.569665 4980 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-6fqt7"] Dec 06 04:01:43 crc kubenswrapper[4980]: I1206 04:01:43.589053 4980 scope.go:117] "RemoveContainer" containerID="3b073f68b1237680b6011a2ec68bb210106ab2eab774d382b7265eec49ca387d" Dec 06 04:01:43 crc kubenswrapper[4980]: I1206 04:01:43.630531 4980 scope.go:117] "RemoveContainer" containerID="80aedde12ae1c998bed1310ca202dd66299e75b51687bcd0b4dba2c3e8fb1aba" Dec 06 04:01:43 crc kubenswrapper[4980]: E1206 04:01:43.631107 4980 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"80aedde12ae1c998bed1310ca202dd66299e75b51687bcd0b4dba2c3e8fb1aba\": container with ID starting with 80aedde12ae1c998bed1310ca202dd66299e75b51687bcd0b4dba2c3e8fb1aba not found: ID does not exist" containerID="80aedde12ae1c998bed1310ca202dd66299e75b51687bcd0b4dba2c3e8fb1aba" Dec 06 04:01:43 crc kubenswrapper[4980]: I1206 04:01:43.631152 4980 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"80aedde12ae1c998bed1310ca202dd66299e75b51687bcd0b4dba2c3e8fb1aba"} err="failed to get container status \"80aedde12ae1c998bed1310ca202dd66299e75b51687bcd0b4dba2c3e8fb1aba\": rpc error: code = NotFound desc = could not find container \"80aedde12ae1c998bed1310ca202dd66299e75b51687bcd0b4dba2c3e8fb1aba\": container with ID starting with 80aedde12ae1c998bed1310ca202dd66299e75b51687bcd0b4dba2c3e8fb1aba not found: ID does not exist" Dec 06 04:01:43 crc kubenswrapper[4980]: I1206 04:01:43.631189 4980 scope.go:117] "RemoveContainer" containerID="52f7c348bfb7fbab86d8644f90533f55ed6f68d5ffe399cb6e89ad6d81066d08" Dec 06 04:01:43 crc kubenswrapper[4980]: E1206 04:01:43.631744 4980 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"52f7c348bfb7fbab86d8644f90533f55ed6f68d5ffe399cb6e89ad6d81066d08\": container with ID starting with 52f7c348bfb7fbab86d8644f90533f55ed6f68d5ffe399cb6e89ad6d81066d08 not found: ID does not exist" containerID="52f7c348bfb7fbab86d8644f90533f55ed6f68d5ffe399cb6e89ad6d81066d08" Dec 06 04:01:43 crc kubenswrapper[4980]: I1206 04:01:43.631767 4980 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"52f7c348bfb7fbab86d8644f90533f55ed6f68d5ffe399cb6e89ad6d81066d08"} err="failed to get container status \"52f7c348bfb7fbab86d8644f90533f55ed6f68d5ffe399cb6e89ad6d81066d08\": rpc error: code = NotFound desc = could not find container \"52f7c348bfb7fbab86d8644f90533f55ed6f68d5ffe399cb6e89ad6d81066d08\": container with ID starting with 52f7c348bfb7fbab86d8644f90533f55ed6f68d5ffe399cb6e89ad6d81066d08 not found: ID does not exist" Dec 06 04:01:43 crc kubenswrapper[4980]: I1206 04:01:43.631784 4980 scope.go:117] "RemoveContainer" containerID="3b073f68b1237680b6011a2ec68bb210106ab2eab774d382b7265eec49ca387d" Dec 06 04:01:43 crc kubenswrapper[4980]: E1206 04:01:43.632056 4980 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3b073f68b1237680b6011a2ec68bb210106ab2eab774d382b7265eec49ca387d\": container with ID starting with 3b073f68b1237680b6011a2ec68bb210106ab2eab774d382b7265eec49ca387d not found: ID does not exist" containerID="3b073f68b1237680b6011a2ec68bb210106ab2eab774d382b7265eec49ca387d" Dec 06 04:01:43 crc kubenswrapper[4980]: I1206 04:01:43.632090 4980 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3b073f68b1237680b6011a2ec68bb210106ab2eab774d382b7265eec49ca387d"} err="failed to get container status \"3b073f68b1237680b6011a2ec68bb210106ab2eab774d382b7265eec49ca387d\": rpc error: code = NotFound desc = could not find container \"3b073f68b1237680b6011a2ec68bb210106ab2eab774d382b7265eec49ca387d\": container with ID starting with 3b073f68b1237680b6011a2ec68bb210106ab2eab774d382b7265eec49ca387d not found: ID does not exist" Dec 06 04:01:45 crc kubenswrapper[4980]: I1206 04:01:45.193797 4980 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9e5ca126-1dff-4106-8643-20657251eeea" path="/var/lib/kubelet/pods/9e5ca126-1dff-4106-8643-20657251eeea/volumes" Dec 06 04:01:51 crc kubenswrapper[4980]: I1206 04:01:51.501530 4980 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-w88cv"] Dec 06 04:01:51 crc kubenswrapper[4980]: E1206 04:01:51.502565 4980 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9e5ca126-1dff-4106-8643-20657251eeea" containerName="extract-content" Dec 06 04:01:51 crc kubenswrapper[4980]: I1206 04:01:51.502587 4980 state_mem.go:107] "Deleted CPUSet assignment" podUID="9e5ca126-1dff-4106-8643-20657251eeea" containerName="extract-content" Dec 06 04:01:51 crc kubenswrapper[4980]: E1206 04:01:51.502614 4980 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9e5ca126-1dff-4106-8643-20657251eeea" containerName="extract-utilities" Dec 06 04:01:51 crc kubenswrapper[4980]: I1206 04:01:51.502628 4980 state_mem.go:107] "Deleted CPUSet assignment" podUID="9e5ca126-1dff-4106-8643-20657251eeea" containerName="extract-utilities" Dec 06 04:01:51 crc kubenswrapper[4980]: E1206 04:01:51.502689 4980 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9e5ca126-1dff-4106-8643-20657251eeea" containerName="registry-server" Dec 06 04:01:51 crc kubenswrapper[4980]: I1206 04:01:51.502701 4980 state_mem.go:107] "Deleted CPUSet assignment" podUID="9e5ca126-1dff-4106-8643-20657251eeea" containerName="registry-server" Dec 06 04:01:51 crc kubenswrapper[4980]: I1206 04:01:51.502937 4980 memory_manager.go:354] "RemoveStaleState removing state" podUID="9e5ca126-1dff-4106-8643-20657251eeea" containerName="registry-server" Dec 06 04:01:51 crc kubenswrapper[4980]: I1206 04:01:51.506116 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-w88cv" Dec 06 04:01:51 crc kubenswrapper[4980]: I1206 04:01:51.522408 4980 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-w88cv"] Dec 06 04:01:51 crc kubenswrapper[4980]: I1206 04:01:51.556966 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/639b5c70-22a0-4b7f-ba5b-72d9cc2b50ad-catalog-content\") pod \"certified-operators-w88cv\" (UID: \"639b5c70-22a0-4b7f-ba5b-72d9cc2b50ad\") " pod="openshift-marketplace/certified-operators-w88cv" Dec 06 04:01:51 crc kubenswrapper[4980]: I1206 04:01:51.557040 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d27xz\" (UniqueName: \"kubernetes.io/projected/639b5c70-22a0-4b7f-ba5b-72d9cc2b50ad-kube-api-access-d27xz\") pod \"certified-operators-w88cv\" (UID: \"639b5c70-22a0-4b7f-ba5b-72d9cc2b50ad\") " pod="openshift-marketplace/certified-operators-w88cv" Dec 06 04:01:51 crc kubenswrapper[4980]: I1206 04:01:51.557134 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/639b5c70-22a0-4b7f-ba5b-72d9cc2b50ad-utilities\") pod \"certified-operators-w88cv\" (UID: \"639b5c70-22a0-4b7f-ba5b-72d9cc2b50ad\") " pod="openshift-marketplace/certified-operators-w88cv" Dec 06 04:01:51 crc kubenswrapper[4980]: I1206 04:01:51.658892 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/639b5c70-22a0-4b7f-ba5b-72d9cc2b50ad-catalog-content\") pod \"certified-operators-w88cv\" (UID: \"639b5c70-22a0-4b7f-ba5b-72d9cc2b50ad\") " pod="openshift-marketplace/certified-operators-w88cv" Dec 06 04:01:51 crc kubenswrapper[4980]: I1206 04:01:51.659304 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d27xz\" (UniqueName: \"kubernetes.io/projected/639b5c70-22a0-4b7f-ba5b-72d9cc2b50ad-kube-api-access-d27xz\") pod \"certified-operators-w88cv\" (UID: \"639b5c70-22a0-4b7f-ba5b-72d9cc2b50ad\") " pod="openshift-marketplace/certified-operators-w88cv" Dec 06 04:01:51 crc kubenswrapper[4980]: I1206 04:01:51.659411 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/639b5c70-22a0-4b7f-ba5b-72d9cc2b50ad-utilities\") pod \"certified-operators-w88cv\" (UID: \"639b5c70-22a0-4b7f-ba5b-72d9cc2b50ad\") " pod="openshift-marketplace/certified-operators-w88cv" Dec 06 04:01:51 crc kubenswrapper[4980]: I1206 04:01:51.659454 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/639b5c70-22a0-4b7f-ba5b-72d9cc2b50ad-catalog-content\") pod \"certified-operators-w88cv\" (UID: \"639b5c70-22a0-4b7f-ba5b-72d9cc2b50ad\") " pod="openshift-marketplace/certified-operators-w88cv" Dec 06 04:01:51 crc kubenswrapper[4980]: I1206 04:01:51.659902 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/639b5c70-22a0-4b7f-ba5b-72d9cc2b50ad-utilities\") pod \"certified-operators-w88cv\" (UID: \"639b5c70-22a0-4b7f-ba5b-72d9cc2b50ad\") " pod="openshift-marketplace/certified-operators-w88cv" Dec 06 04:01:51 crc kubenswrapper[4980]: I1206 04:01:51.680226 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d27xz\" (UniqueName: \"kubernetes.io/projected/639b5c70-22a0-4b7f-ba5b-72d9cc2b50ad-kube-api-access-d27xz\") pod \"certified-operators-w88cv\" (UID: \"639b5c70-22a0-4b7f-ba5b-72d9cc2b50ad\") " pod="openshift-marketplace/certified-operators-w88cv" Dec 06 04:01:51 crc kubenswrapper[4980]: I1206 04:01:51.839368 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-w88cv" Dec 06 04:01:52 crc kubenswrapper[4980]: I1206 04:01:52.184051 4980 scope.go:117] "RemoveContainer" containerID="e8878e13c398cceec46bb52c4c9ea0e06dbf4317e6ad39b69ca763a0d8b15497" Dec 06 04:01:52 crc kubenswrapper[4980]: E1206 04:01:52.184552 4980 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-r5zfb_openshift-machine-config-operator(320f44d1-a671-4a91-b328-a8b0fdd8f23a)\"" pod="openshift-machine-config-operator/machine-config-daemon-r5zfb" podUID="320f44d1-a671-4a91-b328-a8b0fdd8f23a" Dec 06 04:01:52 crc kubenswrapper[4980]: I1206 04:01:52.384349 4980 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-w88cv"] Dec 06 04:01:52 crc kubenswrapper[4980]: I1206 04:01:52.639673 4980 generic.go:334] "Generic (PLEG): container finished" podID="639b5c70-22a0-4b7f-ba5b-72d9cc2b50ad" containerID="2df15b5e44ee646b05cd05e258e39c0777655f39ae6b01f8c35a29a38c309685" exitCode=0 Dec 06 04:01:52 crc kubenswrapper[4980]: I1206 04:01:52.639740 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-w88cv" event={"ID":"639b5c70-22a0-4b7f-ba5b-72d9cc2b50ad","Type":"ContainerDied","Data":"2df15b5e44ee646b05cd05e258e39c0777655f39ae6b01f8c35a29a38c309685"} Dec 06 04:01:52 crc kubenswrapper[4980]: I1206 04:01:52.639781 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-w88cv" event={"ID":"639b5c70-22a0-4b7f-ba5b-72d9cc2b50ad","Type":"ContainerStarted","Data":"9d41a129b14c2d149be5c71741c6a1e8edb49632b37e6a4e8d49a428e3d5e39c"} Dec 06 04:01:52 crc kubenswrapper[4980]: I1206 04:01:52.644399 4980 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 06 04:01:54 crc kubenswrapper[4980]: I1206 04:01:54.664282 4980 generic.go:334] "Generic (PLEG): container finished" podID="639b5c70-22a0-4b7f-ba5b-72d9cc2b50ad" containerID="0e8eb89acc9bd50d3487375127eccd604328dae0ce12231419f0b2dce7ae8947" exitCode=0 Dec 06 04:01:54 crc kubenswrapper[4980]: I1206 04:01:54.664345 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-w88cv" event={"ID":"639b5c70-22a0-4b7f-ba5b-72d9cc2b50ad","Type":"ContainerDied","Data":"0e8eb89acc9bd50d3487375127eccd604328dae0ce12231419f0b2dce7ae8947"} Dec 06 04:01:55 crc kubenswrapper[4980]: I1206 04:01:55.677482 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-w88cv" event={"ID":"639b5c70-22a0-4b7f-ba5b-72d9cc2b50ad","Type":"ContainerStarted","Data":"88d96fe9d049e394cc9b654de7d45ebafcc87228313e61411d2f2199505a0420"} Dec 06 04:01:55 crc kubenswrapper[4980]: I1206 04:01:55.700923 4980 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-w88cv" podStartSLOduration=2.014885255 podStartE2EDuration="4.700877194s" podCreationTimestamp="2025-12-06 04:01:51 +0000 UTC" firstStartedPulling="2025-12-06 04:01:52.643447081 +0000 UTC m=+1671.888198402" lastFinishedPulling="2025-12-06 04:01:55.32943907 +0000 UTC m=+1674.574190341" observedRunningTime="2025-12-06 04:01:55.696036625 +0000 UTC m=+1674.940787916" watchObservedRunningTime="2025-12-06 04:01:55.700877194 +0000 UTC m=+1674.945628475" Dec 06 04:02:01 crc kubenswrapper[4980]: I1206 04:02:01.840122 4980 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-w88cv" Dec 06 04:02:01 crc kubenswrapper[4980]: I1206 04:02:01.840882 4980 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-w88cv" Dec 06 04:02:01 crc kubenswrapper[4980]: I1206 04:02:01.920891 4980 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-w88cv" Dec 06 04:02:02 crc kubenswrapper[4980]: I1206 04:02:02.796619 4980 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-w88cv" Dec 06 04:02:02 crc kubenswrapper[4980]: I1206 04:02:02.863471 4980 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-w88cv"] Dec 06 04:02:04 crc kubenswrapper[4980]: I1206 04:02:04.084620 4980 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/keystone-db-create-t9xwz"] Dec 06 04:02:04 crc kubenswrapper[4980]: I1206 04:02:04.090491 4980 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/keystone-db-create-t9xwz"] Dec 06 04:02:04 crc kubenswrapper[4980]: I1206 04:02:04.760298 4980 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-w88cv" podUID="639b5c70-22a0-4b7f-ba5b-72d9cc2b50ad" containerName="registry-server" containerID="cri-o://88d96fe9d049e394cc9b654de7d45ebafcc87228313e61411d2f2199505a0420" gracePeriod=2 Dec 06 04:02:05 crc kubenswrapper[4980]: I1206 04:02:05.203073 4980 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c88434fa-d55f-4ac5-8075-db953773e17c" path="/var/lib/kubelet/pods/c88434fa-d55f-4ac5-8075-db953773e17c/volumes" Dec 06 04:02:06 crc kubenswrapper[4980]: I1206 04:02:06.247088 4980 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-w88cv" Dec 06 04:02:06 crc kubenswrapper[4980]: I1206 04:02:06.395679 4980 scope.go:117] "RemoveContainer" containerID="fdff1a97bc85df51ddd2b75327e3acbbe68356f8c40a00b7aaa611aac0ccbbc0" Dec 06 04:02:06 crc kubenswrapper[4980]: I1206 04:02:06.399916 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d27xz\" (UniqueName: \"kubernetes.io/projected/639b5c70-22a0-4b7f-ba5b-72d9cc2b50ad-kube-api-access-d27xz\") pod \"639b5c70-22a0-4b7f-ba5b-72d9cc2b50ad\" (UID: \"639b5c70-22a0-4b7f-ba5b-72d9cc2b50ad\") " Dec 06 04:02:06 crc kubenswrapper[4980]: I1206 04:02:06.400048 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/639b5c70-22a0-4b7f-ba5b-72d9cc2b50ad-utilities\") pod \"639b5c70-22a0-4b7f-ba5b-72d9cc2b50ad\" (UID: \"639b5c70-22a0-4b7f-ba5b-72d9cc2b50ad\") " Dec 06 04:02:06 crc kubenswrapper[4980]: I1206 04:02:06.400094 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/639b5c70-22a0-4b7f-ba5b-72d9cc2b50ad-catalog-content\") pod \"639b5c70-22a0-4b7f-ba5b-72d9cc2b50ad\" (UID: \"639b5c70-22a0-4b7f-ba5b-72d9cc2b50ad\") " Dec 06 04:02:06 crc kubenswrapper[4980]: I1206 04:02:06.400947 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/639b5c70-22a0-4b7f-ba5b-72d9cc2b50ad-utilities" (OuterVolumeSpecName: "utilities") pod "639b5c70-22a0-4b7f-ba5b-72d9cc2b50ad" (UID: "639b5c70-22a0-4b7f-ba5b-72d9cc2b50ad"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 04:02:06 crc kubenswrapper[4980]: I1206 04:02:06.412287 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/639b5c70-22a0-4b7f-ba5b-72d9cc2b50ad-kube-api-access-d27xz" (OuterVolumeSpecName: "kube-api-access-d27xz") pod "639b5c70-22a0-4b7f-ba5b-72d9cc2b50ad" (UID: "639b5c70-22a0-4b7f-ba5b-72d9cc2b50ad"). InnerVolumeSpecName "kube-api-access-d27xz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 04:02:06 crc kubenswrapper[4980]: I1206 04:02:06.456330 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/639b5c70-22a0-4b7f-ba5b-72d9cc2b50ad-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "639b5c70-22a0-4b7f-ba5b-72d9cc2b50ad" (UID: "639b5c70-22a0-4b7f-ba5b-72d9cc2b50ad"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 04:02:06 crc kubenswrapper[4980]: I1206 04:02:06.457944 4980 scope.go:117] "RemoveContainer" containerID="e6779bc417ecf996156f1751301fa8bddc16f17f41bdc80c112cb5888d28be19" Dec 06 04:02:06 crc kubenswrapper[4980]: I1206 04:02:06.501641 4980 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d27xz\" (UniqueName: \"kubernetes.io/projected/639b5c70-22a0-4b7f-ba5b-72d9cc2b50ad-kube-api-access-d27xz\") on node \"crc\" DevicePath \"\"" Dec 06 04:02:06 crc kubenswrapper[4980]: I1206 04:02:06.501668 4980 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/639b5c70-22a0-4b7f-ba5b-72d9cc2b50ad-utilities\") on node \"crc\" DevicePath \"\"" Dec 06 04:02:06 crc kubenswrapper[4980]: I1206 04:02:06.501679 4980 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/639b5c70-22a0-4b7f-ba5b-72d9cc2b50ad-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 06 04:02:06 crc kubenswrapper[4980]: I1206 04:02:06.778523 4980 generic.go:334] "Generic (PLEG): container finished" podID="639b5c70-22a0-4b7f-ba5b-72d9cc2b50ad" containerID="88d96fe9d049e394cc9b654de7d45ebafcc87228313e61411d2f2199505a0420" exitCode=0 Dec 06 04:02:06 crc kubenswrapper[4980]: I1206 04:02:06.778579 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-w88cv" event={"ID":"639b5c70-22a0-4b7f-ba5b-72d9cc2b50ad","Type":"ContainerDied","Data":"88d96fe9d049e394cc9b654de7d45ebafcc87228313e61411d2f2199505a0420"} Dec 06 04:02:06 crc kubenswrapper[4980]: I1206 04:02:06.778608 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-w88cv" event={"ID":"639b5c70-22a0-4b7f-ba5b-72d9cc2b50ad","Type":"ContainerDied","Data":"9d41a129b14c2d149be5c71741c6a1e8edb49632b37e6a4e8d49a428e3d5e39c"} Dec 06 04:02:06 crc kubenswrapper[4980]: I1206 04:02:06.778629 4980 scope.go:117] "RemoveContainer" containerID="88d96fe9d049e394cc9b654de7d45ebafcc87228313e61411d2f2199505a0420" Dec 06 04:02:06 crc kubenswrapper[4980]: I1206 04:02:06.778650 4980 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-w88cv" Dec 06 04:02:06 crc kubenswrapper[4980]: I1206 04:02:06.799585 4980 scope.go:117] "RemoveContainer" containerID="0e8eb89acc9bd50d3487375127eccd604328dae0ce12231419f0b2dce7ae8947" Dec 06 04:02:06 crc kubenswrapper[4980]: I1206 04:02:06.828823 4980 scope.go:117] "RemoveContainer" containerID="2df15b5e44ee646b05cd05e258e39c0777655f39ae6b01f8c35a29a38c309685" Dec 06 04:02:06 crc kubenswrapper[4980]: I1206 04:02:06.835575 4980 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-w88cv"] Dec 06 04:02:06 crc kubenswrapper[4980]: I1206 04:02:06.844993 4980 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-w88cv"] Dec 06 04:02:06 crc kubenswrapper[4980]: I1206 04:02:06.850789 4980 scope.go:117] "RemoveContainer" containerID="88d96fe9d049e394cc9b654de7d45ebafcc87228313e61411d2f2199505a0420" Dec 06 04:02:06 crc kubenswrapper[4980]: E1206 04:02:06.851312 4980 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"88d96fe9d049e394cc9b654de7d45ebafcc87228313e61411d2f2199505a0420\": container with ID starting with 88d96fe9d049e394cc9b654de7d45ebafcc87228313e61411d2f2199505a0420 not found: ID does not exist" containerID="88d96fe9d049e394cc9b654de7d45ebafcc87228313e61411d2f2199505a0420" Dec 06 04:02:06 crc kubenswrapper[4980]: I1206 04:02:06.851353 4980 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"88d96fe9d049e394cc9b654de7d45ebafcc87228313e61411d2f2199505a0420"} err="failed to get container status \"88d96fe9d049e394cc9b654de7d45ebafcc87228313e61411d2f2199505a0420\": rpc error: code = NotFound desc = could not find container \"88d96fe9d049e394cc9b654de7d45ebafcc87228313e61411d2f2199505a0420\": container with ID starting with 88d96fe9d049e394cc9b654de7d45ebafcc87228313e61411d2f2199505a0420 not found: ID does not exist" Dec 06 04:02:06 crc kubenswrapper[4980]: I1206 04:02:06.851379 4980 scope.go:117] "RemoveContainer" containerID="0e8eb89acc9bd50d3487375127eccd604328dae0ce12231419f0b2dce7ae8947" Dec 06 04:02:06 crc kubenswrapper[4980]: E1206 04:02:06.852179 4980 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0e8eb89acc9bd50d3487375127eccd604328dae0ce12231419f0b2dce7ae8947\": container with ID starting with 0e8eb89acc9bd50d3487375127eccd604328dae0ce12231419f0b2dce7ae8947 not found: ID does not exist" containerID="0e8eb89acc9bd50d3487375127eccd604328dae0ce12231419f0b2dce7ae8947" Dec 06 04:02:06 crc kubenswrapper[4980]: I1206 04:02:06.852213 4980 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0e8eb89acc9bd50d3487375127eccd604328dae0ce12231419f0b2dce7ae8947"} err="failed to get container status \"0e8eb89acc9bd50d3487375127eccd604328dae0ce12231419f0b2dce7ae8947\": rpc error: code = NotFound desc = could not find container \"0e8eb89acc9bd50d3487375127eccd604328dae0ce12231419f0b2dce7ae8947\": container with ID starting with 0e8eb89acc9bd50d3487375127eccd604328dae0ce12231419f0b2dce7ae8947 not found: ID does not exist" Dec 06 04:02:06 crc kubenswrapper[4980]: I1206 04:02:06.852234 4980 scope.go:117] "RemoveContainer" containerID="2df15b5e44ee646b05cd05e258e39c0777655f39ae6b01f8c35a29a38c309685" Dec 06 04:02:06 crc kubenswrapper[4980]: E1206 04:02:06.852990 4980 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2df15b5e44ee646b05cd05e258e39c0777655f39ae6b01f8c35a29a38c309685\": container with ID starting with 2df15b5e44ee646b05cd05e258e39c0777655f39ae6b01f8c35a29a38c309685 not found: ID does not exist" containerID="2df15b5e44ee646b05cd05e258e39c0777655f39ae6b01f8c35a29a38c309685" Dec 06 04:02:06 crc kubenswrapper[4980]: I1206 04:02:06.853103 4980 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2df15b5e44ee646b05cd05e258e39c0777655f39ae6b01f8c35a29a38c309685"} err="failed to get container status \"2df15b5e44ee646b05cd05e258e39c0777655f39ae6b01f8c35a29a38c309685\": rpc error: code = NotFound desc = could not find container \"2df15b5e44ee646b05cd05e258e39c0777655f39ae6b01f8c35a29a38c309685\": container with ID starting with 2df15b5e44ee646b05cd05e258e39c0777655f39ae6b01f8c35a29a38c309685 not found: ID does not exist" Dec 06 04:02:07 crc kubenswrapper[4980]: I1206 04:02:07.185153 4980 scope.go:117] "RemoveContainer" containerID="e8878e13c398cceec46bb52c4c9ea0e06dbf4317e6ad39b69ca763a0d8b15497" Dec 06 04:02:07 crc kubenswrapper[4980]: E1206 04:02:07.185643 4980 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-r5zfb_openshift-machine-config-operator(320f44d1-a671-4a91-b328-a8b0fdd8f23a)\"" pod="openshift-machine-config-operator/machine-config-daemon-r5zfb" podUID="320f44d1-a671-4a91-b328-a8b0fdd8f23a" Dec 06 04:02:07 crc kubenswrapper[4980]: I1206 04:02:07.197169 4980 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="639b5c70-22a0-4b7f-ba5b-72d9cc2b50ad" path="/var/lib/kubelet/pods/639b5c70-22a0-4b7f-ba5b-72d9cc2b50ad/volumes" Dec 06 04:02:13 crc kubenswrapper[4980]: I1206 04:02:13.053766 4980 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/keystone-0b06-account-create-pf8hh"] Dec 06 04:02:13 crc kubenswrapper[4980]: I1206 04:02:13.069315 4980 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/keystone-0b06-account-create-pf8hh"] Dec 06 04:02:13 crc kubenswrapper[4980]: I1206 04:02:13.200099 4980 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f76dd592-3194-4b0a-81e2-fc87f32f3214" path="/var/lib/kubelet/pods/f76dd592-3194-4b0a-81e2-fc87f32f3214/volumes" Dec 06 04:02:20 crc kubenswrapper[4980]: I1206 04:02:20.184762 4980 scope.go:117] "RemoveContainer" containerID="e8878e13c398cceec46bb52c4c9ea0e06dbf4317e6ad39b69ca763a0d8b15497" Dec 06 04:02:20 crc kubenswrapper[4980]: E1206 04:02:20.185487 4980 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-r5zfb_openshift-machine-config-operator(320f44d1-a671-4a91-b328-a8b0fdd8f23a)\"" pod="openshift-machine-config-operator/machine-config-daemon-r5zfb" podUID="320f44d1-a671-4a91-b328-a8b0fdd8f23a" Dec 06 04:02:30 crc kubenswrapper[4980]: I1206 04:02:30.033313 4980 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/keystone-db-sync-5bvrm"] Dec 06 04:02:30 crc kubenswrapper[4980]: I1206 04:02:30.041710 4980 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/keystone-db-sync-5bvrm"] Dec 06 04:02:31 crc kubenswrapper[4980]: I1206 04:02:31.200115 4980 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d8d8646d-f06f-403c-9ed7-e870564c7329" path="/var/lib/kubelet/pods/d8d8646d-f06f-403c-9ed7-e870564c7329/volumes" Dec 06 04:02:33 crc kubenswrapper[4980]: I1206 04:02:33.184566 4980 scope.go:117] "RemoveContainer" containerID="e8878e13c398cceec46bb52c4c9ea0e06dbf4317e6ad39b69ca763a0d8b15497" Dec 06 04:02:33 crc kubenswrapper[4980]: E1206 04:02:33.184891 4980 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-r5zfb_openshift-machine-config-operator(320f44d1-a671-4a91-b328-a8b0fdd8f23a)\"" pod="openshift-machine-config-operator/machine-config-daemon-r5zfb" podUID="320f44d1-a671-4a91-b328-a8b0fdd8f23a" Dec 06 04:02:37 crc kubenswrapper[4980]: I1206 04:02:37.032216 4980 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/keystone-bootstrap-4lh5q"] Dec 06 04:02:37 crc kubenswrapper[4980]: I1206 04:02:37.038770 4980 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/keystone-bootstrap-4lh5q"] Dec 06 04:02:37 crc kubenswrapper[4980]: I1206 04:02:37.196686 4980 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="01a2d5bc-6b08-4407-b9c4-e588a789cccf" path="/var/lib/kubelet/pods/01a2d5bc-6b08-4407-b9c4-e588a789cccf/volumes" Dec 06 04:02:46 crc kubenswrapper[4980]: I1206 04:02:46.184690 4980 scope.go:117] "RemoveContainer" containerID="e8878e13c398cceec46bb52c4c9ea0e06dbf4317e6ad39b69ca763a0d8b15497" Dec 06 04:02:46 crc kubenswrapper[4980]: E1206 04:02:46.185418 4980 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-r5zfb_openshift-machine-config-operator(320f44d1-a671-4a91-b328-a8b0fdd8f23a)\"" pod="openshift-machine-config-operator/machine-config-daemon-r5zfb" podUID="320f44d1-a671-4a91-b328-a8b0fdd8f23a" Dec 06 04:02:58 crc kubenswrapper[4980]: I1206 04:02:58.184224 4980 scope.go:117] "RemoveContainer" containerID="e8878e13c398cceec46bb52c4c9ea0e06dbf4317e6ad39b69ca763a0d8b15497" Dec 06 04:02:58 crc kubenswrapper[4980]: E1206 04:02:58.184935 4980 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-r5zfb_openshift-machine-config-operator(320f44d1-a671-4a91-b328-a8b0fdd8f23a)\"" pod="openshift-machine-config-operator/machine-config-daemon-r5zfb" podUID="320f44d1-a671-4a91-b328-a8b0fdd8f23a" Dec 06 04:03:06 crc kubenswrapper[4980]: I1206 04:03:06.543874 4980 scope.go:117] "RemoveContainer" containerID="6c4578e685304543174fec33d1b046ae66eb3d061795db4c986db1a947718b35" Dec 06 04:03:06 crc kubenswrapper[4980]: I1206 04:03:06.578007 4980 scope.go:117] "RemoveContainer" containerID="7fec5567a90eb550b62de633c0e2046fb64e4867f89abde77306a213ceb0370f" Dec 06 04:03:06 crc kubenswrapper[4980]: I1206 04:03:06.629773 4980 scope.go:117] "RemoveContainer" containerID="fb73adf57f7fa906fd21886e3d9c5da1211cee117c1e8b663adb5a31071287d8" Dec 06 04:03:06 crc kubenswrapper[4980]: I1206 04:03:06.678401 4980 scope.go:117] "RemoveContainer" containerID="3e4480c29db3d34e5961de14acf7a1876fe97eb74b5a64c421b4e09e932d21ae" Dec 06 04:03:06 crc kubenswrapper[4980]: I1206 04:03:06.701271 4980 scope.go:117] "RemoveContainer" containerID="8951fad3dafe21d4c37c7ed72d986cd9c42fbc4bd08df4cce7c8c43f2e7a4625" Dec 06 04:03:06 crc kubenswrapper[4980]: I1206 04:03:06.725658 4980 scope.go:117] "RemoveContainer" containerID="753730fa2418548cbcb343d5fd27fc6d2f49bbf8d8505e21c4025c5fcd2dc732" Dec 06 04:03:06 crc kubenswrapper[4980]: I1206 04:03:06.749643 4980 scope.go:117] "RemoveContainer" containerID="c451c6ce6fb1fbc8b8c12ae524173dd2515e9eb63d5a7cea60858425ae32143a" Dec 06 04:03:13 crc kubenswrapper[4980]: I1206 04:03:13.184267 4980 scope.go:117] "RemoveContainer" containerID="e8878e13c398cceec46bb52c4c9ea0e06dbf4317e6ad39b69ca763a0d8b15497" Dec 06 04:03:13 crc kubenswrapper[4980]: E1206 04:03:13.185219 4980 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-r5zfb_openshift-machine-config-operator(320f44d1-a671-4a91-b328-a8b0fdd8f23a)\"" pod="openshift-machine-config-operator/machine-config-daemon-r5zfb" podUID="320f44d1-a671-4a91-b328-a8b0fdd8f23a" Dec 06 04:03:24 crc kubenswrapper[4980]: I1206 04:03:24.185003 4980 scope.go:117] "RemoveContainer" containerID="e8878e13c398cceec46bb52c4c9ea0e06dbf4317e6ad39b69ca763a0d8b15497" Dec 06 04:03:24 crc kubenswrapper[4980]: E1206 04:03:24.185905 4980 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-r5zfb_openshift-machine-config-operator(320f44d1-a671-4a91-b328-a8b0fdd8f23a)\"" pod="openshift-machine-config-operator/machine-config-daemon-r5zfb" podUID="320f44d1-a671-4a91-b328-a8b0fdd8f23a" Dec 06 04:03:37 crc kubenswrapper[4980]: I1206 04:03:37.184532 4980 scope.go:117] "RemoveContainer" containerID="e8878e13c398cceec46bb52c4c9ea0e06dbf4317e6ad39b69ca763a0d8b15497" Dec 06 04:03:37 crc kubenswrapper[4980]: E1206 04:03:37.185231 4980 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-r5zfb_openshift-machine-config-operator(320f44d1-a671-4a91-b328-a8b0fdd8f23a)\"" pod="openshift-machine-config-operator/machine-config-daemon-r5zfb" podUID="320f44d1-a671-4a91-b328-a8b0fdd8f23a" Dec 06 04:03:52 crc kubenswrapper[4980]: I1206 04:03:52.184400 4980 scope.go:117] "RemoveContainer" containerID="e8878e13c398cceec46bb52c4c9ea0e06dbf4317e6ad39b69ca763a0d8b15497" Dec 06 04:03:52 crc kubenswrapper[4980]: E1206 04:03:52.185753 4980 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-r5zfb_openshift-machine-config-operator(320f44d1-a671-4a91-b328-a8b0fdd8f23a)\"" pod="openshift-machine-config-operator/machine-config-daemon-r5zfb" podUID="320f44d1-a671-4a91-b328-a8b0fdd8f23a" Dec 06 04:03:53 crc kubenswrapper[4980]: I1206 04:03:53.382158 4980 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/openstackclient"] Dec 06 04:03:53 crc kubenswrapper[4980]: E1206 04:03:53.382728 4980 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="639b5c70-22a0-4b7f-ba5b-72d9cc2b50ad" containerName="registry-server" Dec 06 04:03:53 crc kubenswrapper[4980]: I1206 04:03:53.382740 4980 state_mem.go:107] "Deleted CPUSet assignment" podUID="639b5c70-22a0-4b7f-ba5b-72d9cc2b50ad" containerName="registry-server" Dec 06 04:03:53 crc kubenswrapper[4980]: E1206 04:03:53.382751 4980 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="639b5c70-22a0-4b7f-ba5b-72d9cc2b50ad" containerName="extract-content" Dec 06 04:03:53 crc kubenswrapper[4980]: I1206 04:03:53.382757 4980 state_mem.go:107] "Deleted CPUSet assignment" podUID="639b5c70-22a0-4b7f-ba5b-72d9cc2b50ad" containerName="extract-content" Dec 06 04:03:53 crc kubenswrapper[4980]: E1206 04:03:53.382779 4980 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="639b5c70-22a0-4b7f-ba5b-72d9cc2b50ad" containerName="extract-utilities" Dec 06 04:03:53 crc kubenswrapper[4980]: I1206 04:03:53.382785 4980 state_mem.go:107] "Deleted CPUSet assignment" podUID="639b5c70-22a0-4b7f-ba5b-72d9cc2b50ad" containerName="extract-utilities" Dec 06 04:03:53 crc kubenswrapper[4980]: I1206 04:03:53.382913 4980 memory_manager.go:354] "RemoveStaleState removing state" podUID="639b5c70-22a0-4b7f-ba5b-72d9cc2b50ad" containerName="registry-server" Dec 06 04:03:53 crc kubenswrapper[4980]: I1206 04:03:53.383389 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/openstackclient" Dec 06 04:03:53 crc kubenswrapper[4980]: I1206 04:03:53.385819 4980 reflector.go:368] Caches populated for *v1.ConfigMap from object-"glance-kuttl-tests"/"openstack-config" Dec 06 04:03:53 crc kubenswrapper[4980]: I1206 04:03:53.389465 4980 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"openstack-config-secret" Dec 06 04:03:53 crc kubenswrapper[4980]: I1206 04:03:53.389673 4980 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"default-dockercfg-mc277" Dec 06 04:03:53 crc kubenswrapper[4980]: I1206 04:03:53.390129 4980 reflector.go:368] Caches populated for *v1.ConfigMap from object-"glance-kuttl-tests"/"openstack-scripts-9db6gc427h" Dec 06 04:03:53 crc kubenswrapper[4980]: I1206 04:03:53.396615 4980 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/openstackclient"] Dec 06 04:03:53 crc kubenswrapper[4980]: I1206 04:03:53.541764 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/7ce27058-38ce-4ee9-b71d-b0f7f6dbcb2c-openstack-config\") pod \"openstackclient\" (UID: \"7ce27058-38ce-4ee9-b71d-b0f7f6dbcb2c\") " pod="glance-kuttl-tests/openstackclient" Dec 06 04:03:53 crc kubenswrapper[4980]: I1206 04:03:53.542224 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-scripts\" (UniqueName: \"kubernetes.io/configmap/7ce27058-38ce-4ee9-b71d-b0f7f6dbcb2c-openstack-scripts\") pod \"openstackclient\" (UID: \"7ce27058-38ce-4ee9-b71d-b0f7f6dbcb2c\") " pod="glance-kuttl-tests/openstackclient" Dec 06 04:03:53 crc kubenswrapper[4980]: I1206 04:03:53.542387 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/7ce27058-38ce-4ee9-b71d-b0f7f6dbcb2c-openstack-config-secret\") pod \"openstackclient\" (UID: \"7ce27058-38ce-4ee9-b71d-b0f7f6dbcb2c\") " pod="glance-kuttl-tests/openstackclient" Dec 06 04:03:53 crc kubenswrapper[4980]: I1206 04:03:53.542653 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-92847\" (UniqueName: \"kubernetes.io/projected/7ce27058-38ce-4ee9-b71d-b0f7f6dbcb2c-kube-api-access-92847\") pod \"openstackclient\" (UID: \"7ce27058-38ce-4ee9-b71d-b0f7f6dbcb2c\") " pod="glance-kuttl-tests/openstackclient" Dec 06 04:03:53 crc kubenswrapper[4980]: I1206 04:03:53.644047 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-scripts\" (UniqueName: \"kubernetes.io/configmap/7ce27058-38ce-4ee9-b71d-b0f7f6dbcb2c-openstack-scripts\") pod \"openstackclient\" (UID: \"7ce27058-38ce-4ee9-b71d-b0f7f6dbcb2c\") " pod="glance-kuttl-tests/openstackclient" Dec 06 04:03:53 crc kubenswrapper[4980]: I1206 04:03:53.644114 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/7ce27058-38ce-4ee9-b71d-b0f7f6dbcb2c-openstack-config-secret\") pod \"openstackclient\" (UID: \"7ce27058-38ce-4ee9-b71d-b0f7f6dbcb2c\") " pod="glance-kuttl-tests/openstackclient" Dec 06 04:03:53 crc kubenswrapper[4980]: I1206 04:03:53.644143 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-92847\" (UniqueName: \"kubernetes.io/projected/7ce27058-38ce-4ee9-b71d-b0f7f6dbcb2c-kube-api-access-92847\") pod \"openstackclient\" (UID: \"7ce27058-38ce-4ee9-b71d-b0f7f6dbcb2c\") " pod="glance-kuttl-tests/openstackclient" Dec 06 04:03:53 crc kubenswrapper[4980]: I1206 04:03:53.644225 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/7ce27058-38ce-4ee9-b71d-b0f7f6dbcb2c-openstack-config\") pod \"openstackclient\" (UID: \"7ce27058-38ce-4ee9-b71d-b0f7f6dbcb2c\") " pod="glance-kuttl-tests/openstackclient" Dec 06 04:03:53 crc kubenswrapper[4980]: I1206 04:03:53.645384 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/7ce27058-38ce-4ee9-b71d-b0f7f6dbcb2c-openstack-config\") pod \"openstackclient\" (UID: \"7ce27058-38ce-4ee9-b71d-b0f7f6dbcb2c\") " pod="glance-kuttl-tests/openstackclient" Dec 06 04:03:53 crc kubenswrapper[4980]: I1206 04:03:53.646500 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-scripts\" (UniqueName: \"kubernetes.io/configmap/7ce27058-38ce-4ee9-b71d-b0f7f6dbcb2c-openstack-scripts\") pod \"openstackclient\" (UID: \"7ce27058-38ce-4ee9-b71d-b0f7f6dbcb2c\") " pod="glance-kuttl-tests/openstackclient" Dec 06 04:03:53 crc kubenswrapper[4980]: I1206 04:03:53.651651 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/7ce27058-38ce-4ee9-b71d-b0f7f6dbcb2c-openstack-config-secret\") pod \"openstackclient\" (UID: \"7ce27058-38ce-4ee9-b71d-b0f7f6dbcb2c\") " pod="glance-kuttl-tests/openstackclient" Dec 06 04:03:53 crc kubenswrapper[4980]: I1206 04:03:53.663300 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-92847\" (UniqueName: \"kubernetes.io/projected/7ce27058-38ce-4ee9-b71d-b0f7f6dbcb2c-kube-api-access-92847\") pod \"openstackclient\" (UID: \"7ce27058-38ce-4ee9-b71d-b0f7f6dbcb2c\") " pod="glance-kuttl-tests/openstackclient" Dec 06 04:03:53 crc kubenswrapper[4980]: I1206 04:03:53.707978 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/openstackclient" Dec 06 04:03:54 crc kubenswrapper[4980]: I1206 04:03:54.001927 4980 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/openstackclient"] Dec 06 04:03:54 crc kubenswrapper[4980]: I1206 04:03:54.944795 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/openstackclient" event={"ID":"7ce27058-38ce-4ee9-b71d-b0f7f6dbcb2c","Type":"ContainerStarted","Data":"ce3b9ece9f11a280ab1380d35ad032b941bf6a637401b732e0096602aa8f3f71"} Dec 06 04:03:54 crc kubenswrapper[4980]: I1206 04:03:54.945249 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/openstackclient" event={"ID":"7ce27058-38ce-4ee9-b71d-b0f7f6dbcb2c","Type":"ContainerStarted","Data":"4232d8b41cf662c42cf33e55f92b1205d1000cef7fa78869d5299d9b037c1cbe"} Dec 06 04:03:54 crc kubenswrapper[4980]: I1206 04:03:54.974395 4980 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/openstackclient" podStartSLOduration=1.974351774 podStartE2EDuration="1.974351774s" podCreationTimestamp="2025-12-06 04:03:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 04:03:54.964306697 +0000 UTC m=+1794.209058048" watchObservedRunningTime="2025-12-06 04:03:54.974351774 +0000 UTC m=+1794.219103055" Dec 06 04:04:03 crc kubenswrapper[4980]: I1206 04:04:03.185496 4980 scope.go:117] "RemoveContainer" containerID="e8878e13c398cceec46bb52c4c9ea0e06dbf4317e6ad39b69ca763a0d8b15497" Dec 06 04:04:03 crc kubenswrapper[4980]: E1206 04:04:03.186439 4980 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-r5zfb_openshift-machine-config-operator(320f44d1-a671-4a91-b328-a8b0fdd8f23a)\"" pod="openshift-machine-config-operator/machine-config-daemon-r5zfb" podUID="320f44d1-a671-4a91-b328-a8b0fdd8f23a" Dec 06 04:04:15 crc kubenswrapper[4980]: I1206 04:04:15.184615 4980 scope.go:117] "RemoveContainer" containerID="e8878e13c398cceec46bb52c4c9ea0e06dbf4317e6ad39b69ca763a0d8b15497" Dec 06 04:04:15 crc kubenswrapper[4980]: E1206 04:04:15.185528 4980 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-r5zfb_openshift-machine-config-operator(320f44d1-a671-4a91-b328-a8b0fdd8f23a)\"" pod="openshift-machine-config-operator/machine-config-daemon-r5zfb" podUID="320f44d1-a671-4a91-b328-a8b0fdd8f23a" Dec 06 04:04:28 crc kubenswrapper[4980]: I1206 04:04:28.184312 4980 scope.go:117] "RemoveContainer" containerID="e8878e13c398cceec46bb52c4c9ea0e06dbf4317e6ad39b69ca763a0d8b15497" Dec 06 04:04:29 crc kubenswrapper[4980]: I1206 04:04:29.268889 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-r5zfb" event={"ID":"320f44d1-a671-4a91-b328-a8b0fdd8f23a","Type":"ContainerStarted","Data":"7e5a814e12c0fd9a500ff15c5dc09dfe0e3f68d5cc913a88b3e14e53b9b65070"} Dec 06 04:05:13 crc kubenswrapper[4980]: I1206 04:05:13.328790 4980 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-bgmb7/must-gather-d8t8w"] Dec 06 04:05:13 crc kubenswrapper[4980]: I1206 04:05:13.330471 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-bgmb7/must-gather-d8t8w" Dec 06 04:05:13 crc kubenswrapper[4980]: I1206 04:05:13.332975 4980 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-bgmb7"/"default-dockercfg-6xvxs" Dec 06 04:05:13 crc kubenswrapper[4980]: I1206 04:05:13.333475 4980 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-bgmb7"/"openshift-service-ca.crt" Dec 06 04:05:13 crc kubenswrapper[4980]: I1206 04:05:13.333623 4980 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-bgmb7"/"kube-root-ca.crt" Dec 06 04:05:13 crc kubenswrapper[4980]: I1206 04:05:13.349399 4980 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-bgmb7/must-gather-d8t8w"] Dec 06 04:05:13 crc kubenswrapper[4980]: I1206 04:05:13.506969 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/35cb639f-8637-4bd9-906a-d83988d015ff-must-gather-output\") pod \"must-gather-d8t8w\" (UID: \"35cb639f-8637-4bd9-906a-d83988d015ff\") " pod="openshift-must-gather-bgmb7/must-gather-d8t8w" Dec 06 04:05:13 crc kubenswrapper[4980]: I1206 04:05:13.507282 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tpr6q\" (UniqueName: \"kubernetes.io/projected/35cb639f-8637-4bd9-906a-d83988d015ff-kube-api-access-tpr6q\") pod \"must-gather-d8t8w\" (UID: \"35cb639f-8637-4bd9-906a-d83988d015ff\") " pod="openshift-must-gather-bgmb7/must-gather-d8t8w" Dec 06 04:05:13 crc kubenswrapper[4980]: I1206 04:05:13.609319 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/35cb639f-8637-4bd9-906a-d83988d015ff-must-gather-output\") pod \"must-gather-d8t8w\" (UID: \"35cb639f-8637-4bd9-906a-d83988d015ff\") " pod="openshift-must-gather-bgmb7/must-gather-d8t8w" Dec 06 04:05:13 crc kubenswrapper[4980]: I1206 04:05:13.609414 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tpr6q\" (UniqueName: \"kubernetes.io/projected/35cb639f-8637-4bd9-906a-d83988d015ff-kube-api-access-tpr6q\") pod \"must-gather-d8t8w\" (UID: \"35cb639f-8637-4bd9-906a-d83988d015ff\") " pod="openshift-must-gather-bgmb7/must-gather-d8t8w" Dec 06 04:05:13 crc kubenswrapper[4980]: I1206 04:05:13.609866 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/35cb639f-8637-4bd9-906a-d83988d015ff-must-gather-output\") pod \"must-gather-d8t8w\" (UID: \"35cb639f-8637-4bd9-906a-d83988d015ff\") " pod="openshift-must-gather-bgmb7/must-gather-d8t8w" Dec 06 04:05:13 crc kubenswrapper[4980]: I1206 04:05:13.634463 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tpr6q\" (UniqueName: \"kubernetes.io/projected/35cb639f-8637-4bd9-906a-d83988d015ff-kube-api-access-tpr6q\") pod \"must-gather-d8t8w\" (UID: \"35cb639f-8637-4bd9-906a-d83988d015ff\") " pod="openshift-must-gather-bgmb7/must-gather-d8t8w" Dec 06 04:05:13 crc kubenswrapper[4980]: I1206 04:05:13.652631 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-bgmb7/must-gather-d8t8w" Dec 06 04:05:14 crc kubenswrapper[4980]: I1206 04:05:14.074924 4980 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-bgmb7/must-gather-d8t8w"] Dec 06 04:05:14 crc kubenswrapper[4980]: I1206 04:05:14.659492 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-bgmb7/must-gather-d8t8w" event={"ID":"35cb639f-8637-4bd9-906a-d83988d015ff","Type":"ContainerStarted","Data":"4fda628e73c9401e397de66f5d2fde52fd9e1b12755bd8002d32c20b27ffc9de"} Dec 06 04:05:18 crc kubenswrapper[4980]: I1206 04:05:18.713620 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-bgmb7/must-gather-d8t8w" event={"ID":"35cb639f-8637-4bd9-906a-d83988d015ff","Type":"ContainerStarted","Data":"cf7175d2595ea07c5997befacc560760e8ab20ff7b7e2a469abcaad88333d487"} Dec 06 04:05:18 crc kubenswrapper[4980]: I1206 04:05:18.715348 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-bgmb7/must-gather-d8t8w" event={"ID":"35cb639f-8637-4bd9-906a-d83988d015ff","Type":"ContainerStarted","Data":"8c0c956e04fbd4dad778a309ea2058c2348a4e65773bf8bb813e89a67e4fe405"} Dec 06 04:05:18 crc kubenswrapper[4980]: I1206 04:05:18.730400 4980 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-bgmb7/must-gather-d8t8w" podStartSLOduration=2.01886729 podStartE2EDuration="5.730374204s" podCreationTimestamp="2025-12-06 04:05:13 +0000 UTC" firstStartedPulling="2025-12-06 04:05:14.091253971 +0000 UTC m=+1873.336005242" lastFinishedPulling="2025-12-06 04:05:17.802760885 +0000 UTC m=+1877.047512156" observedRunningTime="2025-12-06 04:05:18.727210394 +0000 UTC m=+1877.971961655" watchObservedRunningTime="2025-12-06 04:05:18.730374204 +0000 UTC m=+1877.975125465" Dec 06 04:05:54 crc kubenswrapper[4980]: I1206 04:05:54.918409 4980 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_4d9ed08db2179c6d86e46659534222f77f18c6c6d13f1f155004bdcf89l5n8z_6674a828-934e-4b14-98be-740dad507b95/util/0.log" Dec 06 04:05:55 crc kubenswrapper[4980]: I1206 04:05:55.051593 4980 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_4d9ed08db2179c6d86e46659534222f77f18c6c6d13f1f155004bdcf89l5n8z_6674a828-934e-4b14-98be-740dad507b95/util/0.log" Dec 06 04:05:55 crc kubenswrapper[4980]: I1206 04:05:55.053941 4980 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_4d9ed08db2179c6d86e46659534222f77f18c6c6d13f1f155004bdcf89l5n8z_6674a828-934e-4b14-98be-740dad507b95/pull/0.log" Dec 06 04:05:55 crc kubenswrapper[4980]: I1206 04:05:55.105487 4980 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_4d9ed08db2179c6d86e46659534222f77f18c6c6d13f1f155004bdcf89l5n8z_6674a828-934e-4b14-98be-740dad507b95/pull/0.log" Dec 06 04:05:55 crc kubenswrapper[4980]: I1206 04:05:55.250576 4980 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_4d9ed08db2179c6d86e46659534222f77f18c6c6d13f1f155004bdcf89l5n8z_6674a828-934e-4b14-98be-740dad507b95/pull/0.log" Dec 06 04:05:55 crc kubenswrapper[4980]: I1206 04:05:55.252144 4980 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_4d9ed08db2179c6d86e46659534222f77f18c6c6d13f1f155004bdcf89l5n8z_6674a828-934e-4b14-98be-740dad507b95/util/0.log" Dec 06 04:05:55 crc kubenswrapper[4980]: I1206 04:05:55.294660 4980 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_4d9ed08db2179c6d86e46659534222f77f18c6c6d13f1f155004bdcf89l5n8z_6674a828-934e-4b14-98be-740dad507b95/extract/0.log" Dec 06 04:05:55 crc kubenswrapper[4980]: I1206 04:05:55.431879 4980 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_673e18c374f427d47ebf1299aef3a4ea1aab2302d0e987986af30e26212ztjs_7171eff1-10af-4bab-b43b-02368608ee00/util/0.log" Dec 06 04:05:55 crc kubenswrapper[4980]: I1206 04:05:55.597122 4980 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_673e18c374f427d47ebf1299aef3a4ea1aab2302d0e987986af30e26212ztjs_7171eff1-10af-4bab-b43b-02368608ee00/pull/0.log" Dec 06 04:05:55 crc kubenswrapper[4980]: I1206 04:05:55.614066 4980 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_673e18c374f427d47ebf1299aef3a4ea1aab2302d0e987986af30e26212ztjs_7171eff1-10af-4bab-b43b-02368608ee00/pull/0.log" Dec 06 04:05:55 crc kubenswrapper[4980]: I1206 04:05:55.634619 4980 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_673e18c374f427d47ebf1299aef3a4ea1aab2302d0e987986af30e26212ztjs_7171eff1-10af-4bab-b43b-02368608ee00/util/0.log" Dec 06 04:05:55 crc kubenswrapper[4980]: I1206 04:05:55.787405 4980 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_673e18c374f427d47ebf1299aef3a4ea1aab2302d0e987986af30e26212ztjs_7171eff1-10af-4bab-b43b-02368608ee00/pull/0.log" Dec 06 04:05:55 crc kubenswrapper[4980]: I1206 04:05:55.811678 4980 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_673e18c374f427d47ebf1299aef3a4ea1aab2302d0e987986af30e26212ztjs_7171eff1-10af-4bab-b43b-02368608ee00/util/0.log" Dec 06 04:05:55 crc kubenswrapper[4980]: I1206 04:05:55.828126 4980 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_673e18c374f427d47ebf1299aef3a4ea1aab2302d0e987986af30e26212ztjs_7171eff1-10af-4bab-b43b-02368608ee00/extract/0.log" Dec 06 04:05:55 crc kubenswrapper[4980]: I1206 04:05:55.967470 4980 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_798531b9d9a078af26f5f153dd8093f0980ac32bb052a41050c010ef74b88wp_33b3d1b9-c7e6-44e3-9259-1251e4e340a6/util/0.log" Dec 06 04:05:56 crc kubenswrapper[4980]: I1206 04:05:56.160563 4980 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_798531b9d9a078af26f5f153dd8093f0980ac32bb052a41050c010ef74b88wp_33b3d1b9-c7e6-44e3-9259-1251e4e340a6/pull/0.log" Dec 06 04:05:56 crc kubenswrapper[4980]: I1206 04:05:56.185290 4980 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_798531b9d9a078af26f5f153dd8093f0980ac32bb052a41050c010ef74b88wp_33b3d1b9-c7e6-44e3-9259-1251e4e340a6/pull/0.log" Dec 06 04:05:56 crc kubenswrapper[4980]: I1206 04:05:56.211294 4980 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_798531b9d9a078af26f5f153dd8093f0980ac32bb052a41050c010ef74b88wp_33b3d1b9-c7e6-44e3-9259-1251e4e340a6/util/0.log" Dec 06 04:05:56 crc kubenswrapper[4980]: I1206 04:05:56.322392 4980 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_798531b9d9a078af26f5f153dd8093f0980ac32bb052a41050c010ef74b88wp_33b3d1b9-c7e6-44e3-9259-1251e4e340a6/pull/0.log" Dec 06 04:05:56 crc kubenswrapper[4980]: I1206 04:05:56.326728 4980 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_798531b9d9a078af26f5f153dd8093f0980ac32bb052a41050c010ef74b88wp_33b3d1b9-c7e6-44e3-9259-1251e4e340a6/util/0.log" Dec 06 04:05:56 crc kubenswrapper[4980]: I1206 04:05:56.381682 4980 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_798531b9d9a078af26f5f153dd8093f0980ac32bb052a41050c010ef74b88wp_33b3d1b9-c7e6-44e3-9259-1251e4e340a6/extract/0.log" Dec 06 04:05:56 crc kubenswrapper[4980]: I1206 04:05:56.495497 4980 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_82cf678a38d30dcf5cc31c16a6d10360b45b3df2f5eaffd238882752a25cbw6_37352474-24e2-4b44-b5d7-c482f8cffa3f/util/0.log" Dec 06 04:05:56 crc kubenswrapper[4980]: I1206 04:05:56.663653 4980 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_82cf678a38d30dcf5cc31c16a6d10360b45b3df2f5eaffd238882752a25cbw6_37352474-24e2-4b44-b5d7-c482f8cffa3f/pull/0.log" Dec 06 04:05:56 crc kubenswrapper[4980]: I1206 04:05:56.665209 4980 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_82cf678a38d30dcf5cc31c16a6d10360b45b3df2f5eaffd238882752a25cbw6_37352474-24e2-4b44-b5d7-c482f8cffa3f/util/0.log" Dec 06 04:05:56 crc kubenswrapper[4980]: I1206 04:05:56.694072 4980 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_82cf678a38d30dcf5cc31c16a6d10360b45b3df2f5eaffd238882752a25cbw6_37352474-24e2-4b44-b5d7-c482f8cffa3f/pull/0.log" Dec 06 04:05:56 crc kubenswrapper[4980]: I1206 04:05:56.903707 4980 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_82cf678a38d30dcf5cc31c16a6d10360b45b3df2f5eaffd238882752a25cbw6_37352474-24e2-4b44-b5d7-c482f8cffa3f/extract/0.log" Dec 06 04:05:56 crc kubenswrapper[4980]: I1206 04:05:56.906551 4980 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_82cf678a38d30dcf5cc31c16a6d10360b45b3df2f5eaffd238882752a25cbw6_37352474-24e2-4b44-b5d7-c482f8cffa3f/util/0.log" Dec 06 04:05:56 crc kubenswrapper[4980]: I1206 04:05:56.908083 4980 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_82cf678a38d30dcf5cc31c16a6d10360b45b3df2f5eaffd238882752a25cbw6_37352474-24e2-4b44-b5d7-c482f8cffa3f/pull/0.log" Dec 06 04:05:57 crc kubenswrapper[4980]: I1206 04:05:57.107366 4980 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_8ad853bb090b3a36d4332bde14850afdc9872f13e5d95ff50094430eb62kjwk_a8036768-6216-4f7a-adc6-7f81287574e3/util/0.log" Dec 06 04:05:57 crc kubenswrapper[4980]: I1206 04:05:57.416289 4980 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_8ad853bb090b3a36d4332bde14850afdc9872f13e5d95ff50094430eb62kjwk_a8036768-6216-4f7a-adc6-7f81287574e3/pull/0.log" Dec 06 04:05:57 crc kubenswrapper[4980]: I1206 04:05:57.422544 4980 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_8ad853bb090b3a36d4332bde14850afdc9872f13e5d95ff50094430eb62kjwk_a8036768-6216-4f7a-adc6-7f81287574e3/util/0.log" Dec 06 04:05:57 crc kubenswrapper[4980]: I1206 04:05:57.456202 4980 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_8ad853bb090b3a36d4332bde14850afdc9872f13e5d95ff50094430eb62kjwk_a8036768-6216-4f7a-adc6-7f81287574e3/pull/0.log" Dec 06 04:05:57 crc kubenswrapper[4980]: I1206 04:05:57.634814 4980 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_8ad853bb090b3a36d4332bde14850afdc9872f13e5d95ff50094430eb62kjwk_a8036768-6216-4f7a-adc6-7f81287574e3/util/0.log" Dec 06 04:05:57 crc kubenswrapper[4980]: I1206 04:05:57.668759 4980 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_8ad853bb090b3a36d4332bde14850afdc9872f13e5d95ff50094430eb62kjwk_a8036768-6216-4f7a-adc6-7f81287574e3/extract/0.log" Dec 06 04:05:57 crc kubenswrapper[4980]: I1206 04:05:57.678020 4980 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_8ad853bb090b3a36d4332bde14850afdc9872f13e5d95ff50094430eb62kjwk_a8036768-6216-4f7a-adc6-7f81287574e3/pull/0.log" Dec 06 04:05:57 crc kubenswrapper[4980]: I1206 04:05:57.825833 4980 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590cnjz7_b2cb8251-b0fb-4d95-a2fb-460ab17081e4/util/0.log" Dec 06 04:05:57 crc kubenswrapper[4980]: I1206 04:05:57.944368 4980 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590cnjz7_b2cb8251-b0fb-4d95-a2fb-460ab17081e4/pull/0.log" Dec 06 04:05:57 crc kubenswrapper[4980]: I1206 04:05:57.953375 4980 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590cnjz7_b2cb8251-b0fb-4d95-a2fb-460ab17081e4/pull/0.log" Dec 06 04:05:57 crc kubenswrapper[4980]: I1206 04:05:57.955419 4980 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590cnjz7_b2cb8251-b0fb-4d95-a2fb-460ab17081e4/util/0.log" Dec 06 04:05:58 crc kubenswrapper[4980]: I1206 04:05:58.172677 4980 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590cnjz7_b2cb8251-b0fb-4d95-a2fb-460ab17081e4/pull/0.log" Dec 06 04:05:58 crc kubenswrapper[4980]: I1206 04:05:58.181958 4980 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590cnjz7_b2cb8251-b0fb-4d95-a2fb-460ab17081e4/util/0.log" Dec 06 04:05:58 crc kubenswrapper[4980]: I1206 04:05:58.197064 4980 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590cnjz7_b2cb8251-b0fb-4d95-a2fb-460ab17081e4/extract/0.log" Dec 06 04:05:58 crc kubenswrapper[4980]: I1206 04:05:58.199066 4980 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_c03115aeb6cae4480b1000a58090a522fd8798c774d18bf66e63c4d3d15vssx_b59d13fa-5f2a-4f67-9502-4b81c035ee10/util/0.log" Dec 06 04:05:58 crc kubenswrapper[4980]: I1206 04:05:58.439749 4980 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_c03115aeb6cae4480b1000a58090a522fd8798c774d18bf66e63c4d3d15vssx_b59d13fa-5f2a-4f67-9502-4b81c035ee10/util/0.log" Dec 06 04:05:58 crc kubenswrapper[4980]: I1206 04:05:58.457234 4980 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_c03115aeb6cae4480b1000a58090a522fd8798c774d18bf66e63c4d3d15vssx_b59d13fa-5f2a-4f67-9502-4b81c035ee10/pull/0.log" Dec 06 04:05:58 crc kubenswrapper[4980]: I1206 04:05:58.460170 4980 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_c03115aeb6cae4480b1000a58090a522fd8798c774d18bf66e63c4d3d15vssx_b59d13fa-5f2a-4f67-9502-4b81c035ee10/pull/0.log" Dec 06 04:05:58 crc kubenswrapper[4980]: I1206 04:05:58.625216 4980 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_c03115aeb6cae4480b1000a58090a522fd8798c774d18bf66e63c4d3d15vssx_b59d13fa-5f2a-4f67-9502-4b81c035ee10/util/0.log" Dec 06 04:05:58 crc kubenswrapper[4980]: I1206 04:05:58.626986 4980 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_c03115aeb6cae4480b1000a58090a522fd8798c774d18bf66e63c4d3d15vssx_b59d13fa-5f2a-4f67-9502-4b81c035ee10/extract/0.log" Dec 06 04:05:58 crc kubenswrapper[4980]: I1206 04:05:58.656955 4980 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_c03115aeb6cae4480b1000a58090a522fd8798c774d18bf66e63c4d3d15vssx_b59d13fa-5f2a-4f67-9502-4b81c035ee10/pull/0.log" Dec 06 04:05:58 crc kubenswrapper[4980]: I1206 04:05:58.700618 4980 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-549766cbfc-nxztk_c28fde7c-003e-45ed-bb75-fec17db60f64/kube-rbac-proxy/0.log" Dec 06 04:05:58 crc kubenswrapper[4980]: I1206 04:05:58.877132 4980 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-549766cbfc-nxztk_c28fde7c-003e-45ed-bb75-fec17db60f64/manager/0.log" Dec 06 04:05:58 crc kubenswrapper[4980]: I1206 04:05:58.886761 4980 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-index-bzpqg_c88f8da4-819c-42bf-9938-60c898c6bace/registry-server/0.log" Dec 06 04:05:58 crc kubenswrapper[4980]: I1206 04:05:58.902757 4980 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-558977579-kknvn_c4a8121a-cce4-481a-b758-ae1b5dd89336/kube-rbac-proxy/0.log" Dec 06 04:05:59 crc kubenswrapper[4980]: I1206 04:05:59.072855 4980 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-index-8jmnt_24052d23-b7d3-436b-906f-dcdc363733f9/registry-server/0.log" Dec 06 04:05:59 crc kubenswrapper[4980]: I1206 04:05:59.082525 4980 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-558977579-kknvn_c4a8121a-cce4-481a-b758-ae1b5dd89336/manager/0.log" Dec 06 04:05:59 crc kubenswrapper[4980]: I1206 04:05:59.101938 4980 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-99587f5d7-rcrhn_0cc75ddc-77ef-4c34-b7d1-c5e98f7666dc/kube-rbac-proxy/0.log" Dec 06 04:05:59 crc kubenswrapper[4980]: I1206 04:05:59.262356 4980 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-99587f5d7-rcrhn_0cc75ddc-77ef-4c34-b7d1-c5e98f7666dc/manager/0.log" Dec 06 04:05:59 crc kubenswrapper[4980]: I1206 04:05:59.312195 4980 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-index-gwm9t_8999e732-bb13-4416-b63e-c54355c9f489/registry-server/0.log" Dec 06 04:05:59 crc kubenswrapper[4980]: I1206 04:05:59.346275 4980 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-8485c4fcb7-9qgh9_3f51fcb0-cd67-4d0d-8731-48e926391011/kube-rbac-proxy/0.log" Dec 06 04:05:59 crc kubenswrapper[4980]: I1206 04:05:59.466529 4980 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-8485c4fcb7-9qgh9_3f51fcb0-cd67-4d0d-8731-48e926391011/manager/0.log" Dec 06 04:05:59 crc kubenswrapper[4980]: I1206 04:05:59.557732 4980 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-f567f77d7-bg9qv_710bbcb6-d1c3-46a3-9f04-82ca8e5af2fc/kube-rbac-proxy/0.log" Dec 06 04:05:59 crc kubenswrapper[4980]: I1206 04:05:59.571542 4980 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-index-4jvwv_209b6261-c681-433d-884b-6edad0dc505d/registry-server/0.log" Dec 06 04:05:59 crc kubenswrapper[4980]: I1206 04:05:59.647902 4980 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-f567f77d7-bg9qv_710bbcb6-d1c3-46a3-9f04-82ca8e5af2fc/manager/0.log" Dec 06 04:05:59 crc kubenswrapper[4980]: I1206 04:05:59.712179 4980 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-index-5wm7n_4066d752-9e8f-47d6-a553-f7f7e7c06e25/registry-server/0.log" Dec 06 04:05:59 crc kubenswrapper[4980]: I1206 04:05:59.837104 4980 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_rabbitmq-cluster-operator-779fc9694b-r7tt8_003f4f15-503c-4321-9daa-46f4ca2541d1/operator/0.log" Dec 06 04:05:59 crc kubenswrapper[4980]: I1206 04:05:59.954734 4980 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_rabbitmq-cluster-operator-index-7m77w_b514140e-e078-40e1-b380-5c097450eeac/registry-server/0.log" Dec 06 04:06:00 crc kubenswrapper[4980]: I1206 04:06:00.012705 4980 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-5d8ccb4767-r2zdb_81b1b1bc-6f44-4bca-9b34-87c42f15c07e/kube-rbac-proxy/0.log" Dec 06 04:06:00 crc kubenswrapper[4980]: I1206 04:06:00.028067 4980 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-5d8ccb4767-r2zdb_81b1b1bc-6f44-4bca-9b34-87c42f15c07e/manager/0.log" Dec 06 04:06:00 crc kubenswrapper[4980]: I1206 04:06:00.183053 4980 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-index-rxptj_713b6c54-f0c9-4c60-9f98-0c2e57dae276/registry-server/0.log" Dec 06 04:06:15 crc kubenswrapper[4980]: I1206 04:06:15.079709 4980 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-cqk6j_e81d9a6d-991f-46b0-aba5-6f76a5b1031a/control-plane-machine-set-operator/0.log" Dec 06 04:06:15 crc kubenswrapper[4980]: I1206 04:06:15.245896 4980 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-x5m4p_f8973f62-160b-4c4b-9238-6a2ff3d63bc5/kube-rbac-proxy/0.log" Dec 06 04:06:15 crc kubenswrapper[4980]: I1206 04:06:15.277427 4980 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-x5m4p_f8973f62-160b-4c4b-9238-6a2ff3d63bc5/machine-api-operator/0.log" Dec 06 04:06:27 crc kubenswrapper[4980]: I1206 04:06:27.112874 4980 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-dn7wc"] Dec 06 04:06:27 crc kubenswrapper[4980]: I1206 04:06:27.123378 4980 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-dn7wc"] Dec 06 04:06:27 crc kubenswrapper[4980]: I1206 04:06:27.123580 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-dn7wc" Dec 06 04:06:27 crc kubenswrapper[4980]: I1206 04:06:27.273903 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/adebfc1e-1c83-4e98-8cbb-e6ac2876cbf7-utilities\") pod \"redhat-marketplace-dn7wc\" (UID: \"adebfc1e-1c83-4e98-8cbb-e6ac2876cbf7\") " pod="openshift-marketplace/redhat-marketplace-dn7wc" Dec 06 04:06:27 crc kubenswrapper[4980]: I1206 04:06:27.274756 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rfpj9\" (UniqueName: \"kubernetes.io/projected/adebfc1e-1c83-4e98-8cbb-e6ac2876cbf7-kube-api-access-rfpj9\") pod \"redhat-marketplace-dn7wc\" (UID: \"adebfc1e-1c83-4e98-8cbb-e6ac2876cbf7\") " pod="openshift-marketplace/redhat-marketplace-dn7wc" Dec 06 04:06:27 crc kubenswrapper[4980]: I1206 04:06:27.274988 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/adebfc1e-1c83-4e98-8cbb-e6ac2876cbf7-catalog-content\") pod \"redhat-marketplace-dn7wc\" (UID: \"adebfc1e-1c83-4e98-8cbb-e6ac2876cbf7\") " pod="openshift-marketplace/redhat-marketplace-dn7wc" Dec 06 04:06:27 crc kubenswrapper[4980]: I1206 04:06:27.376615 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/adebfc1e-1c83-4e98-8cbb-e6ac2876cbf7-utilities\") pod \"redhat-marketplace-dn7wc\" (UID: \"adebfc1e-1c83-4e98-8cbb-e6ac2876cbf7\") " pod="openshift-marketplace/redhat-marketplace-dn7wc" Dec 06 04:06:27 crc kubenswrapper[4980]: I1206 04:06:27.377576 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rfpj9\" (UniqueName: \"kubernetes.io/projected/adebfc1e-1c83-4e98-8cbb-e6ac2876cbf7-kube-api-access-rfpj9\") pod \"redhat-marketplace-dn7wc\" (UID: \"adebfc1e-1c83-4e98-8cbb-e6ac2876cbf7\") " pod="openshift-marketplace/redhat-marketplace-dn7wc" Dec 06 04:06:27 crc kubenswrapper[4980]: I1206 04:06:27.377625 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/adebfc1e-1c83-4e98-8cbb-e6ac2876cbf7-catalog-content\") pod \"redhat-marketplace-dn7wc\" (UID: \"adebfc1e-1c83-4e98-8cbb-e6ac2876cbf7\") " pod="openshift-marketplace/redhat-marketplace-dn7wc" Dec 06 04:06:27 crc kubenswrapper[4980]: I1206 04:06:27.378154 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/adebfc1e-1c83-4e98-8cbb-e6ac2876cbf7-catalog-content\") pod \"redhat-marketplace-dn7wc\" (UID: \"adebfc1e-1c83-4e98-8cbb-e6ac2876cbf7\") " pod="openshift-marketplace/redhat-marketplace-dn7wc" Dec 06 04:06:27 crc kubenswrapper[4980]: I1206 04:06:27.378440 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/adebfc1e-1c83-4e98-8cbb-e6ac2876cbf7-utilities\") pod \"redhat-marketplace-dn7wc\" (UID: \"adebfc1e-1c83-4e98-8cbb-e6ac2876cbf7\") " pod="openshift-marketplace/redhat-marketplace-dn7wc" Dec 06 04:06:27 crc kubenswrapper[4980]: I1206 04:06:27.401999 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rfpj9\" (UniqueName: \"kubernetes.io/projected/adebfc1e-1c83-4e98-8cbb-e6ac2876cbf7-kube-api-access-rfpj9\") pod \"redhat-marketplace-dn7wc\" (UID: \"adebfc1e-1c83-4e98-8cbb-e6ac2876cbf7\") " pod="openshift-marketplace/redhat-marketplace-dn7wc" Dec 06 04:06:27 crc kubenswrapper[4980]: I1206 04:06:27.456588 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-dn7wc" Dec 06 04:06:27 crc kubenswrapper[4980]: I1206 04:06:27.917055 4980 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-dn7wc"] Dec 06 04:06:28 crc kubenswrapper[4980]: I1206 04:06:28.299910 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dn7wc" event={"ID":"adebfc1e-1c83-4e98-8cbb-e6ac2876cbf7","Type":"ContainerStarted","Data":"0f574ea43670cba954101fb05ff215d601c147f73be01b889608d558f5ee1c0f"} Dec 06 04:06:29 crc kubenswrapper[4980]: I1206 04:06:29.312715 4980 generic.go:334] "Generic (PLEG): container finished" podID="adebfc1e-1c83-4e98-8cbb-e6ac2876cbf7" containerID="9d2406e5f5be67f00c1d08e81aae6f90145086613eb64617686433d773d7f7e0" exitCode=0 Dec 06 04:06:29 crc kubenswrapper[4980]: I1206 04:06:29.312814 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dn7wc" event={"ID":"adebfc1e-1c83-4e98-8cbb-e6ac2876cbf7","Type":"ContainerDied","Data":"9d2406e5f5be67f00c1d08e81aae6f90145086613eb64617686433d773d7f7e0"} Dec 06 04:06:30 crc kubenswrapper[4980]: I1206 04:06:30.325028 4980 generic.go:334] "Generic (PLEG): container finished" podID="adebfc1e-1c83-4e98-8cbb-e6ac2876cbf7" containerID="1901770f64afe7764431c1046f8d0df5d6f0faf6d0cb7277d96689a17503526d" exitCode=0 Dec 06 04:06:30 crc kubenswrapper[4980]: I1206 04:06:30.325110 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dn7wc" event={"ID":"adebfc1e-1c83-4e98-8cbb-e6ac2876cbf7","Type":"ContainerDied","Data":"1901770f64afe7764431c1046f8d0df5d6f0faf6d0cb7277d96689a17503526d"} Dec 06 04:06:31 crc kubenswrapper[4980]: I1206 04:06:31.334937 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dn7wc" event={"ID":"adebfc1e-1c83-4e98-8cbb-e6ac2876cbf7","Type":"ContainerStarted","Data":"41105d6bac13d23be9ec7f0436c06983781f7867db1fda2d17e677d32cde6373"} Dec 06 04:06:31 crc kubenswrapper[4980]: I1206 04:06:31.363389 4980 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-dn7wc" podStartSLOduration=2.74599657 podStartE2EDuration="4.363350074s" podCreationTimestamp="2025-12-06 04:06:27 +0000 UTC" firstStartedPulling="2025-12-06 04:06:29.314232564 +0000 UTC m=+1948.558983835" lastFinishedPulling="2025-12-06 04:06:30.931586058 +0000 UTC m=+1950.176337339" observedRunningTime="2025-12-06 04:06:31.35794145 +0000 UTC m=+1950.602692741" watchObservedRunningTime="2025-12-06 04:06:31.363350074 +0000 UTC m=+1950.608101345" Dec 06 04:06:32 crc kubenswrapper[4980]: I1206 04:06:32.292750 4980 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-f8648f98b-jg82p_a8cfafe9-920a-4bdd-a857-dae21346a492/kube-rbac-proxy/0.log" Dec 06 04:06:32 crc kubenswrapper[4980]: I1206 04:06:32.371797 4980 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-f8648f98b-jg82p_a8cfafe9-920a-4bdd-a857-dae21346a492/controller/0.log" Dec 06 04:06:32 crc kubenswrapper[4980]: I1206 04:06:32.585486 4980 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-jh6gj_97b725b0-8d1d-4c85-8c96-f96e269da5c5/cp-frr-files/0.log" Dec 06 04:06:32 crc kubenswrapper[4980]: I1206 04:06:32.763335 4980 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-jh6gj_97b725b0-8d1d-4c85-8c96-f96e269da5c5/cp-reloader/0.log" Dec 06 04:06:32 crc kubenswrapper[4980]: I1206 04:06:32.769570 4980 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-jh6gj_97b725b0-8d1d-4c85-8c96-f96e269da5c5/cp-metrics/0.log" Dec 06 04:06:32 crc kubenswrapper[4980]: I1206 04:06:32.773751 4980 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-jh6gj_97b725b0-8d1d-4c85-8c96-f96e269da5c5/cp-frr-files/0.log" Dec 06 04:06:32 crc kubenswrapper[4980]: I1206 04:06:32.800576 4980 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-jh6gj_97b725b0-8d1d-4c85-8c96-f96e269da5c5/cp-reloader/0.log" Dec 06 04:06:32 crc kubenswrapper[4980]: I1206 04:06:32.967382 4980 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-jh6gj_97b725b0-8d1d-4c85-8c96-f96e269da5c5/cp-reloader/0.log" Dec 06 04:06:32 crc kubenswrapper[4980]: I1206 04:06:32.967923 4980 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-jh6gj_97b725b0-8d1d-4c85-8c96-f96e269da5c5/cp-metrics/0.log" Dec 06 04:06:32 crc kubenswrapper[4980]: I1206 04:06:32.969777 4980 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-jh6gj_97b725b0-8d1d-4c85-8c96-f96e269da5c5/cp-frr-files/0.log" Dec 06 04:06:33 crc kubenswrapper[4980]: I1206 04:06:33.014952 4980 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-jh6gj_97b725b0-8d1d-4c85-8c96-f96e269da5c5/cp-metrics/0.log" Dec 06 04:06:33 crc kubenswrapper[4980]: I1206 04:06:33.156720 4980 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-jh6gj_97b725b0-8d1d-4c85-8c96-f96e269da5c5/cp-frr-files/0.log" Dec 06 04:06:33 crc kubenswrapper[4980]: I1206 04:06:33.164504 4980 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-jh6gj_97b725b0-8d1d-4c85-8c96-f96e269da5c5/cp-metrics/0.log" Dec 06 04:06:33 crc kubenswrapper[4980]: I1206 04:06:33.165575 4980 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-jh6gj_97b725b0-8d1d-4c85-8c96-f96e269da5c5/cp-reloader/0.log" Dec 06 04:06:33 crc kubenswrapper[4980]: I1206 04:06:33.234935 4980 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-jh6gj_97b725b0-8d1d-4c85-8c96-f96e269da5c5/controller/0.log" Dec 06 04:06:33 crc kubenswrapper[4980]: I1206 04:06:33.344617 4980 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-jh6gj_97b725b0-8d1d-4c85-8c96-f96e269da5c5/frr-metrics/0.log" Dec 06 04:06:33 crc kubenswrapper[4980]: I1206 04:06:33.369555 4980 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-jh6gj_97b725b0-8d1d-4c85-8c96-f96e269da5c5/kube-rbac-proxy/0.log" Dec 06 04:06:33 crc kubenswrapper[4980]: I1206 04:06:33.439274 4980 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-jh6gj_97b725b0-8d1d-4c85-8c96-f96e269da5c5/kube-rbac-proxy-frr/0.log" Dec 06 04:06:33 crc kubenswrapper[4980]: I1206 04:06:33.550551 4980 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-jh6gj_97b725b0-8d1d-4c85-8c96-f96e269da5c5/reloader/0.log" Dec 06 04:06:33 crc kubenswrapper[4980]: I1206 04:06:33.669959 4980 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-webhook-server-7fcb986d4-7qtqr_1f860039-c5ba-4ba9-90b1-ecb00110015f/frr-k8s-webhook-server/0.log" Dec 06 04:06:33 crc kubenswrapper[4980]: I1206 04:06:33.789886 4980 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-jh6gj_97b725b0-8d1d-4c85-8c96-f96e269da5c5/frr/0.log" Dec 06 04:06:33 crc kubenswrapper[4980]: I1206 04:06:33.838726 4980 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-controller-manager-555b967656-x878b_245c6ddf-2c16-419a-9931-f5887917d145/manager/0.log" Dec 06 04:06:33 crc kubenswrapper[4980]: I1206 04:06:33.980440 4980 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-webhook-server-86db4b77c-4tjd7_7e314ba1-bb63-4da4-b43c-97f9bc74b99f/webhook-server/0.log" Dec 06 04:06:34 crc kubenswrapper[4980]: I1206 04:06:34.062662 4980 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-6wv7f_53351890-4d11-40c9-91c9-b91c24578d70/kube-rbac-proxy/0.log" Dec 06 04:06:34 crc kubenswrapper[4980]: I1206 04:06:34.198036 4980 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-6wv7f_53351890-4d11-40c9-91c9-b91c24578d70/speaker/0.log" Dec 06 04:06:37 crc kubenswrapper[4980]: I1206 04:06:37.457688 4980 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-dn7wc" Dec 06 04:06:37 crc kubenswrapper[4980]: I1206 04:06:37.457742 4980 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-dn7wc" Dec 06 04:06:37 crc kubenswrapper[4980]: I1206 04:06:37.499377 4980 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-dn7wc" Dec 06 04:06:38 crc kubenswrapper[4980]: I1206 04:06:38.473813 4980 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-dn7wc" Dec 06 04:06:38 crc kubenswrapper[4980]: I1206 04:06:38.533012 4980 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-dn7wc"] Dec 06 04:06:40 crc kubenswrapper[4980]: I1206 04:06:40.448007 4980 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-dn7wc" podUID="adebfc1e-1c83-4e98-8cbb-e6ac2876cbf7" containerName="registry-server" containerID="cri-o://41105d6bac13d23be9ec7f0436c06983781f7867db1fda2d17e677d32cde6373" gracePeriod=2 Dec 06 04:06:41 crc kubenswrapper[4980]: I1206 04:06:41.351956 4980 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-dn7wc" Dec 06 04:06:41 crc kubenswrapper[4980]: I1206 04:06:41.452324 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/adebfc1e-1c83-4e98-8cbb-e6ac2876cbf7-utilities\") pod \"adebfc1e-1c83-4e98-8cbb-e6ac2876cbf7\" (UID: \"adebfc1e-1c83-4e98-8cbb-e6ac2876cbf7\") " Dec 06 04:06:41 crc kubenswrapper[4980]: I1206 04:06:41.452639 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rfpj9\" (UniqueName: \"kubernetes.io/projected/adebfc1e-1c83-4e98-8cbb-e6ac2876cbf7-kube-api-access-rfpj9\") pod \"adebfc1e-1c83-4e98-8cbb-e6ac2876cbf7\" (UID: \"adebfc1e-1c83-4e98-8cbb-e6ac2876cbf7\") " Dec 06 04:06:41 crc kubenswrapper[4980]: I1206 04:06:41.452800 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/adebfc1e-1c83-4e98-8cbb-e6ac2876cbf7-catalog-content\") pod \"adebfc1e-1c83-4e98-8cbb-e6ac2876cbf7\" (UID: \"adebfc1e-1c83-4e98-8cbb-e6ac2876cbf7\") " Dec 06 04:06:41 crc kubenswrapper[4980]: I1206 04:06:41.453234 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/adebfc1e-1c83-4e98-8cbb-e6ac2876cbf7-utilities" (OuterVolumeSpecName: "utilities") pod "adebfc1e-1c83-4e98-8cbb-e6ac2876cbf7" (UID: "adebfc1e-1c83-4e98-8cbb-e6ac2876cbf7"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 04:06:41 crc kubenswrapper[4980]: I1206 04:06:41.457471 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/adebfc1e-1c83-4e98-8cbb-e6ac2876cbf7-kube-api-access-rfpj9" (OuterVolumeSpecName: "kube-api-access-rfpj9") pod "adebfc1e-1c83-4e98-8cbb-e6ac2876cbf7" (UID: "adebfc1e-1c83-4e98-8cbb-e6ac2876cbf7"). InnerVolumeSpecName "kube-api-access-rfpj9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 04:06:41 crc kubenswrapper[4980]: I1206 04:06:41.458824 4980 generic.go:334] "Generic (PLEG): container finished" podID="adebfc1e-1c83-4e98-8cbb-e6ac2876cbf7" containerID="41105d6bac13d23be9ec7f0436c06983781f7867db1fda2d17e677d32cde6373" exitCode=0 Dec 06 04:06:41 crc kubenswrapper[4980]: I1206 04:06:41.458861 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dn7wc" event={"ID":"adebfc1e-1c83-4e98-8cbb-e6ac2876cbf7","Type":"ContainerDied","Data":"41105d6bac13d23be9ec7f0436c06983781f7867db1fda2d17e677d32cde6373"} Dec 06 04:06:41 crc kubenswrapper[4980]: I1206 04:06:41.458896 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dn7wc" event={"ID":"adebfc1e-1c83-4e98-8cbb-e6ac2876cbf7","Type":"ContainerDied","Data":"0f574ea43670cba954101fb05ff215d601c147f73be01b889608d558f5ee1c0f"} Dec 06 04:06:41 crc kubenswrapper[4980]: I1206 04:06:41.458929 4980 scope.go:117] "RemoveContainer" containerID="41105d6bac13d23be9ec7f0436c06983781f7867db1fda2d17e677d32cde6373" Dec 06 04:06:41 crc kubenswrapper[4980]: I1206 04:06:41.458953 4980 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-dn7wc" Dec 06 04:06:41 crc kubenswrapper[4980]: I1206 04:06:41.477871 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/adebfc1e-1c83-4e98-8cbb-e6ac2876cbf7-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "adebfc1e-1c83-4e98-8cbb-e6ac2876cbf7" (UID: "adebfc1e-1c83-4e98-8cbb-e6ac2876cbf7"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 04:06:41 crc kubenswrapper[4980]: I1206 04:06:41.523709 4980 scope.go:117] "RemoveContainer" containerID="1901770f64afe7764431c1046f8d0df5d6f0faf6d0cb7277d96689a17503526d" Dec 06 04:06:41 crc kubenswrapper[4980]: I1206 04:06:41.543706 4980 scope.go:117] "RemoveContainer" containerID="9d2406e5f5be67f00c1d08e81aae6f90145086613eb64617686433d773d7f7e0" Dec 06 04:06:41 crc kubenswrapper[4980]: I1206 04:06:41.554247 4980 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/adebfc1e-1c83-4e98-8cbb-e6ac2876cbf7-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 06 04:06:41 crc kubenswrapper[4980]: I1206 04:06:41.554280 4980 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/adebfc1e-1c83-4e98-8cbb-e6ac2876cbf7-utilities\") on node \"crc\" DevicePath \"\"" Dec 06 04:06:41 crc kubenswrapper[4980]: I1206 04:06:41.554290 4980 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rfpj9\" (UniqueName: \"kubernetes.io/projected/adebfc1e-1c83-4e98-8cbb-e6ac2876cbf7-kube-api-access-rfpj9\") on node \"crc\" DevicePath \"\"" Dec 06 04:06:41 crc kubenswrapper[4980]: I1206 04:06:41.575452 4980 scope.go:117] "RemoveContainer" containerID="41105d6bac13d23be9ec7f0436c06983781f7867db1fda2d17e677d32cde6373" Dec 06 04:06:41 crc kubenswrapper[4980]: E1206 04:06:41.575981 4980 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"41105d6bac13d23be9ec7f0436c06983781f7867db1fda2d17e677d32cde6373\": container with ID starting with 41105d6bac13d23be9ec7f0436c06983781f7867db1fda2d17e677d32cde6373 not found: ID does not exist" containerID="41105d6bac13d23be9ec7f0436c06983781f7867db1fda2d17e677d32cde6373" Dec 06 04:06:41 crc kubenswrapper[4980]: I1206 04:06:41.576015 4980 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"41105d6bac13d23be9ec7f0436c06983781f7867db1fda2d17e677d32cde6373"} err="failed to get container status \"41105d6bac13d23be9ec7f0436c06983781f7867db1fda2d17e677d32cde6373\": rpc error: code = NotFound desc = could not find container \"41105d6bac13d23be9ec7f0436c06983781f7867db1fda2d17e677d32cde6373\": container with ID starting with 41105d6bac13d23be9ec7f0436c06983781f7867db1fda2d17e677d32cde6373 not found: ID does not exist" Dec 06 04:06:41 crc kubenswrapper[4980]: I1206 04:06:41.576037 4980 scope.go:117] "RemoveContainer" containerID="1901770f64afe7764431c1046f8d0df5d6f0faf6d0cb7277d96689a17503526d" Dec 06 04:06:41 crc kubenswrapper[4980]: E1206 04:06:41.576573 4980 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1901770f64afe7764431c1046f8d0df5d6f0faf6d0cb7277d96689a17503526d\": container with ID starting with 1901770f64afe7764431c1046f8d0df5d6f0faf6d0cb7277d96689a17503526d not found: ID does not exist" containerID="1901770f64afe7764431c1046f8d0df5d6f0faf6d0cb7277d96689a17503526d" Dec 06 04:06:41 crc kubenswrapper[4980]: I1206 04:06:41.576594 4980 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1901770f64afe7764431c1046f8d0df5d6f0faf6d0cb7277d96689a17503526d"} err="failed to get container status \"1901770f64afe7764431c1046f8d0df5d6f0faf6d0cb7277d96689a17503526d\": rpc error: code = NotFound desc = could not find container \"1901770f64afe7764431c1046f8d0df5d6f0faf6d0cb7277d96689a17503526d\": container with ID starting with 1901770f64afe7764431c1046f8d0df5d6f0faf6d0cb7277d96689a17503526d not found: ID does not exist" Dec 06 04:06:41 crc kubenswrapper[4980]: I1206 04:06:41.576609 4980 scope.go:117] "RemoveContainer" containerID="9d2406e5f5be67f00c1d08e81aae6f90145086613eb64617686433d773d7f7e0" Dec 06 04:06:41 crc kubenswrapper[4980]: E1206 04:06:41.576975 4980 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9d2406e5f5be67f00c1d08e81aae6f90145086613eb64617686433d773d7f7e0\": container with ID starting with 9d2406e5f5be67f00c1d08e81aae6f90145086613eb64617686433d773d7f7e0 not found: ID does not exist" containerID="9d2406e5f5be67f00c1d08e81aae6f90145086613eb64617686433d773d7f7e0" Dec 06 04:06:41 crc kubenswrapper[4980]: I1206 04:06:41.577003 4980 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9d2406e5f5be67f00c1d08e81aae6f90145086613eb64617686433d773d7f7e0"} err="failed to get container status \"9d2406e5f5be67f00c1d08e81aae6f90145086613eb64617686433d773d7f7e0\": rpc error: code = NotFound desc = could not find container \"9d2406e5f5be67f00c1d08e81aae6f90145086613eb64617686433d773d7f7e0\": container with ID starting with 9d2406e5f5be67f00c1d08e81aae6f90145086613eb64617686433d773d7f7e0 not found: ID does not exist" Dec 06 04:06:41 crc kubenswrapper[4980]: I1206 04:06:41.792824 4980 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-dn7wc"] Dec 06 04:06:41 crc kubenswrapper[4980]: I1206 04:06:41.798532 4980 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-dn7wc"] Dec 06 04:06:43 crc kubenswrapper[4980]: I1206 04:06:43.214310 4980 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="adebfc1e-1c83-4e98-8cbb-e6ac2876cbf7" path="/var/lib/kubelet/pods/adebfc1e-1c83-4e98-8cbb-e6ac2876cbf7/volumes" Dec 06 04:06:49 crc kubenswrapper[4980]: I1206 04:06:49.207927 4980 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_glance-8c70-account-create-mkrvh_aae1a12f-6f92-43cc-99ae-fd14a894aadd/mariadb-account-create/0.log" Dec 06 04:06:49 crc kubenswrapper[4980]: I1206 04:06:49.286451 4980 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_glance-cache-glance-default-external-api-0-cleaner-2941656ml2pt_735c2ee4-228e-498d-babc-b395ba133918/glance-cache-glance-default-external-api-0-cleaner/0.log" Dec 06 04:06:49 crc kubenswrapper[4980]: I1206 04:06:49.420811 4980 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_glance-cache-glance-default-internal-api-0-cleaner-294165644rpt_659983bd-c2de-4f2f-88a4-395f494156c8/glance-cache-glance-default-internal-api-0-cleaner/0.log" Dec 06 04:06:49 crc kubenswrapper[4980]: I1206 04:06:49.456561 4980 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_glance-db-create-4wjqb_d9f133d0-65f5-4ee3-9906-97e1a883ddf8/mariadb-database-create/0.log" Dec 06 04:06:49 crc kubenswrapper[4980]: I1206 04:06:49.697540 4980 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_glance-db-sync-9lbg6_34f14e45-81a8-490f-a760-88fbac6f06b3/glance-db-sync/0.log" Dec 06 04:06:49 crc kubenswrapper[4980]: I1206 04:06:49.763306 4980 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_glance-default-external-api-0_7d66b977-3b53-4fe4-bc08-9e9083b32d76/glance-api/0.log" Dec 06 04:06:49 crc kubenswrapper[4980]: I1206 04:06:49.898926 4980 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_glance-default-external-api-0_7d66b977-3b53-4fe4-bc08-9e9083b32d76/glance-httpd/0.log" Dec 06 04:06:49 crc kubenswrapper[4980]: I1206 04:06:49.917379 4980 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_glance-default-external-api-0_7d66b977-3b53-4fe4-bc08-9e9083b32d76/glance-log/0.log" Dec 06 04:06:49 crc kubenswrapper[4980]: I1206 04:06:49.955426 4980 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_glance-default-internal-api-0_3ef73fa3-e8eb-4b78-9831-7e3e0f24bdb8/glance-api/0.log" Dec 06 04:06:50 crc kubenswrapper[4980]: I1206 04:06:50.093387 4980 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_glance-default-internal-api-0_3ef73fa3-e8eb-4b78-9831-7e3e0f24bdb8/glance-httpd/0.log" Dec 06 04:06:50 crc kubenswrapper[4980]: I1206 04:06:50.108567 4980 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_glance-default-internal-api-0_3ef73fa3-e8eb-4b78-9831-7e3e0f24bdb8/glance-log/0.log" Dec 06 04:06:50 crc kubenswrapper[4980]: I1206 04:06:50.278837 4980 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_keystone-cron-29416561-bcltb_786dd383-8c53-46e3-afc3-eebbf6ecae33/keystone-cron/0.log" Dec 06 04:06:50 crc kubenswrapper[4980]: I1206 04:06:50.401977 4980 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_keystone-568cc6458b-84xg6_cd336afa-48b9-4df5-9cca-12031ce13b2d/keystone-api/0.log" Dec 06 04:06:50 crc kubenswrapper[4980]: I1206 04:06:50.614838 4980 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_openstack-galera-0_7267968c-6dd4-44a8-ad87-ab5d0663d88a/mysql-bootstrap/0.log" Dec 06 04:06:50 crc kubenswrapper[4980]: I1206 04:06:50.899182 4980 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_openstack-galera-0_7267968c-6dd4-44a8-ad87-ab5d0663d88a/mysql-bootstrap/0.log" Dec 06 04:06:50 crc kubenswrapper[4980]: I1206 04:06:50.923630 4980 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_openstack-galera-0_7267968c-6dd4-44a8-ad87-ab5d0663d88a/galera/0.log" Dec 06 04:06:51 crc kubenswrapper[4980]: I1206 04:06:51.140841 4980 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_openstack-galera-1_1f395fef-da33-429d-b7e0-61573b301d8d/mysql-bootstrap/0.log" Dec 06 04:06:51 crc kubenswrapper[4980]: I1206 04:06:51.268426 4980 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_openstack-galera-1_1f395fef-da33-429d-b7e0-61573b301d8d/mysql-bootstrap/0.log" Dec 06 04:06:51 crc kubenswrapper[4980]: I1206 04:06:51.375252 4980 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_openstack-galera-1_1f395fef-da33-429d-b7e0-61573b301d8d/galera/0.log" Dec 06 04:06:51 crc kubenswrapper[4980]: I1206 04:06:51.545338 4980 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_openstack-galera-2_c53b1fad-da29-4e9c-970c-9f4d84693905/mysql-bootstrap/0.log" Dec 06 04:06:51 crc kubenswrapper[4980]: I1206 04:06:51.777354 4980 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_openstack-galera-2_c53b1fad-da29-4e9c-970c-9f4d84693905/mysql-bootstrap/0.log" Dec 06 04:06:51 crc kubenswrapper[4980]: I1206 04:06:51.805667 4980 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_openstack-galera-2_c53b1fad-da29-4e9c-970c-9f4d84693905/galera/0.log" Dec 06 04:06:51 crc kubenswrapper[4980]: I1206 04:06:51.959449 4980 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_memcached-0_cbbe1b5d-848b-4c0f-bd83-10b27906dbfc/memcached/0.log" Dec 06 04:06:52 crc kubenswrapper[4980]: I1206 04:06:52.032289 4980 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_rabbitmq-server-0_92922f24-43cc-445d-a4ad-adcbf68b6636/setup-container/0.log" Dec 06 04:06:52 crc kubenswrapper[4980]: I1206 04:06:52.037953 4980 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_openstackclient_7ce27058-38ce-4ee9-b71d-b0f7f6dbcb2c/openstackclient/0.log" Dec 06 04:06:52 crc kubenswrapper[4980]: I1206 04:06:52.165228 4980 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_rabbitmq-server-0_92922f24-43cc-445d-a4ad-adcbf68b6636/setup-container/0.log" Dec 06 04:06:52 crc kubenswrapper[4980]: I1206 04:06:52.237195 4980 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_rabbitmq-server-0_92922f24-43cc-445d-a4ad-adcbf68b6636/rabbitmq/0.log" Dec 06 04:06:52 crc kubenswrapper[4980]: I1206 04:06:52.287387 4980 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_swift-proxy-85fcf7dd9-sn4m2_92027099-00cc-45db-bbfe-7796fe5a68c4/proxy-httpd/0.log" Dec 06 04:06:52 crc kubenswrapper[4980]: I1206 04:06:52.402916 4980 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_swift-ring-rebalance-pfjgb_43cddc9e-ad31-4596-8e89-cb3260fd55a8/swift-ring-rebalance/0.log" Dec 06 04:06:52 crc kubenswrapper[4980]: I1206 04:06:52.448796 4980 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_swift-proxy-85fcf7dd9-sn4m2_92027099-00cc-45db-bbfe-7796fe5a68c4/proxy-server/0.log" Dec 06 04:06:52 crc kubenswrapper[4980]: I1206 04:06:52.509296 4980 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_swift-storage-0_3752191f-1260-4be2-8d68-308491d20468/account-auditor/0.log" Dec 06 04:06:52 crc kubenswrapper[4980]: I1206 04:06:52.595311 4980 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_swift-storage-0_3752191f-1260-4be2-8d68-308491d20468/account-reaper/0.log" Dec 06 04:06:52 crc kubenswrapper[4980]: I1206 04:06:52.616776 4980 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_swift-storage-0_3752191f-1260-4be2-8d68-308491d20468/account-replicator/0.log" Dec 06 04:06:52 crc kubenswrapper[4980]: I1206 04:06:52.663288 4980 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_swift-storage-0_3752191f-1260-4be2-8d68-308491d20468/account-server/0.log" Dec 06 04:06:52 crc kubenswrapper[4980]: I1206 04:06:52.680873 4980 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_swift-storage-0_3752191f-1260-4be2-8d68-308491d20468/container-replicator/0.log" Dec 06 04:06:52 crc kubenswrapper[4980]: I1206 04:06:52.705813 4980 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_swift-storage-0_3752191f-1260-4be2-8d68-308491d20468/container-auditor/0.log" Dec 06 04:06:52 crc kubenswrapper[4980]: I1206 04:06:52.789801 4980 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_swift-storage-0_3752191f-1260-4be2-8d68-308491d20468/container-updater/0.log" Dec 06 04:06:52 crc kubenswrapper[4980]: I1206 04:06:52.790216 4980 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_swift-storage-0_3752191f-1260-4be2-8d68-308491d20468/container-server/0.log" Dec 06 04:06:52 crc kubenswrapper[4980]: I1206 04:06:52.853737 4980 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_swift-storage-0_3752191f-1260-4be2-8d68-308491d20468/object-auditor/0.log" Dec 06 04:06:52 crc kubenswrapper[4980]: I1206 04:06:52.862044 4980 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_swift-storage-0_3752191f-1260-4be2-8d68-308491d20468/object-expirer/0.log" Dec 06 04:06:52 crc kubenswrapper[4980]: I1206 04:06:52.899973 4980 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_swift-storage-0_3752191f-1260-4be2-8d68-308491d20468/object-replicator/0.log" Dec 06 04:06:53 crc kubenswrapper[4980]: I1206 04:06:53.000531 4980 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_swift-storage-0_3752191f-1260-4be2-8d68-308491d20468/object-updater/0.log" Dec 06 04:06:53 crc kubenswrapper[4980]: I1206 04:06:53.013187 4980 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_swift-storage-0_3752191f-1260-4be2-8d68-308491d20468/object-server/0.log" Dec 06 04:06:53 crc kubenswrapper[4980]: I1206 04:06:53.064606 4980 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_swift-storage-0_3752191f-1260-4be2-8d68-308491d20468/rsync/0.log" Dec 06 04:06:53 crc kubenswrapper[4980]: I1206 04:06:53.106408 4980 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_swift-storage-0_3752191f-1260-4be2-8d68-308491d20468/swift-recon-cron/0.log" Dec 06 04:06:55 crc kubenswrapper[4980]: I1206 04:06:55.137445 4980 patch_prober.go:28] interesting pod/machine-config-daemon-r5zfb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 04:06:55 crc kubenswrapper[4980]: I1206 04:06:55.137583 4980 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-r5zfb" podUID="320f44d1-a671-4a91-b328-a8b0fdd8f23a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 04:07:06 crc kubenswrapper[4980]: I1206 04:07:06.762903 4980 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-tgs44"] Dec 06 04:07:06 crc kubenswrapper[4980]: E1206 04:07:06.763997 4980 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="adebfc1e-1c83-4e98-8cbb-e6ac2876cbf7" containerName="extract-utilities" Dec 06 04:07:06 crc kubenswrapper[4980]: I1206 04:07:06.764024 4980 state_mem.go:107] "Deleted CPUSet assignment" podUID="adebfc1e-1c83-4e98-8cbb-e6ac2876cbf7" containerName="extract-utilities" Dec 06 04:07:06 crc kubenswrapper[4980]: E1206 04:07:06.764055 4980 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="adebfc1e-1c83-4e98-8cbb-e6ac2876cbf7" containerName="extract-content" Dec 06 04:07:06 crc kubenswrapper[4980]: I1206 04:07:06.764064 4980 state_mem.go:107] "Deleted CPUSet assignment" podUID="adebfc1e-1c83-4e98-8cbb-e6ac2876cbf7" containerName="extract-content" Dec 06 04:07:06 crc kubenswrapper[4980]: E1206 04:07:06.764084 4980 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="adebfc1e-1c83-4e98-8cbb-e6ac2876cbf7" containerName="registry-server" Dec 06 04:07:06 crc kubenswrapper[4980]: I1206 04:07:06.764093 4980 state_mem.go:107] "Deleted CPUSet assignment" podUID="adebfc1e-1c83-4e98-8cbb-e6ac2876cbf7" containerName="registry-server" Dec 06 04:07:06 crc kubenswrapper[4980]: I1206 04:07:06.764300 4980 memory_manager.go:354] "RemoveStaleState removing state" podUID="adebfc1e-1c83-4e98-8cbb-e6ac2876cbf7" containerName="registry-server" Dec 06 04:07:06 crc kubenswrapper[4980]: I1206 04:07:06.766085 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-tgs44" Dec 06 04:07:06 crc kubenswrapper[4980]: I1206 04:07:06.774290 4980 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-tgs44"] Dec 06 04:07:06 crc kubenswrapper[4980]: I1206 04:07:06.815507 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cd6724f3-6ebc-4c26-9d2c-eaa0de6c503c-utilities\") pod \"redhat-operators-tgs44\" (UID: \"cd6724f3-6ebc-4c26-9d2c-eaa0de6c503c\") " pod="openshift-marketplace/redhat-operators-tgs44" Dec 06 04:07:06 crc kubenswrapper[4980]: I1206 04:07:06.816093 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cd6724f3-6ebc-4c26-9d2c-eaa0de6c503c-catalog-content\") pod \"redhat-operators-tgs44\" (UID: \"cd6724f3-6ebc-4c26-9d2c-eaa0de6c503c\") " pod="openshift-marketplace/redhat-operators-tgs44" Dec 06 04:07:06 crc kubenswrapper[4980]: I1206 04:07:06.816263 4980 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qf897\" (UniqueName: \"kubernetes.io/projected/cd6724f3-6ebc-4c26-9d2c-eaa0de6c503c-kube-api-access-qf897\") pod \"redhat-operators-tgs44\" (UID: \"cd6724f3-6ebc-4c26-9d2c-eaa0de6c503c\") " pod="openshift-marketplace/redhat-operators-tgs44" Dec 06 04:07:06 crc kubenswrapper[4980]: I1206 04:07:06.917616 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cd6724f3-6ebc-4c26-9d2c-eaa0de6c503c-utilities\") pod \"redhat-operators-tgs44\" (UID: \"cd6724f3-6ebc-4c26-9d2c-eaa0de6c503c\") " pod="openshift-marketplace/redhat-operators-tgs44" Dec 06 04:07:06 crc kubenswrapper[4980]: I1206 04:07:06.917737 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cd6724f3-6ebc-4c26-9d2c-eaa0de6c503c-catalog-content\") pod \"redhat-operators-tgs44\" (UID: \"cd6724f3-6ebc-4c26-9d2c-eaa0de6c503c\") " pod="openshift-marketplace/redhat-operators-tgs44" Dec 06 04:07:06 crc kubenswrapper[4980]: I1206 04:07:06.917793 4980 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qf897\" (UniqueName: \"kubernetes.io/projected/cd6724f3-6ebc-4c26-9d2c-eaa0de6c503c-kube-api-access-qf897\") pod \"redhat-operators-tgs44\" (UID: \"cd6724f3-6ebc-4c26-9d2c-eaa0de6c503c\") " pod="openshift-marketplace/redhat-operators-tgs44" Dec 06 04:07:06 crc kubenswrapper[4980]: I1206 04:07:06.918491 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cd6724f3-6ebc-4c26-9d2c-eaa0de6c503c-utilities\") pod \"redhat-operators-tgs44\" (UID: \"cd6724f3-6ebc-4c26-9d2c-eaa0de6c503c\") " pod="openshift-marketplace/redhat-operators-tgs44" Dec 06 04:07:06 crc kubenswrapper[4980]: I1206 04:07:06.929070 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cd6724f3-6ebc-4c26-9d2c-eaa0de6c503c-catalog-content\") pod \"redhat-operators-tgs44\" (UID: \"cd6724f3-6ebc-4c26-9d2c-eaa0de6c503c\") " pod="openshift-marketplace/redhat-operators-tgs44" Dec 06 04:07:06 crc kubenswrapper[4980]: I1206 04:07:06.959947 4980 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qf897\" (UniqueName: \"kubernetes.io/projected/cd6724f3-6ebc-4c26-9d2c-eaa0de6c503c-kube-api-access-qf897\") pod \"redhat-operators-tgs44\" (UID: \"cd6724f3-6ebc-4c26-9d2c-eaa0de6c503c\") " pod="openshift-marketplace/redhat-operators-tgs44" Dec 06 04:07:07 crc kubenswrapper[4980]: I1206 04:07:07.105790 4980 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-tgs44" Dec 06 04:07:07 crc kubenswrapper[4980]: I1206 04:07:07.552897 4980 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83brxgn_73b1ef5a-140c-40df-a801-dc06cc56b5e2/util/0.log" Dec 06 04:07:07 crc kubenswrapper[4980]: I1206 04:07:07.597746 4980 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-tgs44"] Dec 06 04:07:07 crc kubenswrapper[4980]: W1206 04:07:07.615914 4980 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podcd6724f3_6ebc_4c26_9d2c_eaa0de6c503c.slice/crio-9b513ecfadfbd0765ffaaa5bf1a43810790553ea796377c81b4deff0a9555415 WatchSource:0}: Error finding container 9b513ecfadfbd0765ffaaa5bf1a43810790553ea796377c81b4deff0a9555415: Status 404 returned error can't find the container with id 9b513ecfadfbd0765ffaaa5bf1a43810790553ea796377c81b4deff0a9555415 Dec 06 04:07:07 crc kubenswrapper[4980]: I1206 04:07:07.714476 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-tgs44" event={"ID":"cd6724f3-6ebc-4c26-9d2c-eaa0de6c503c","Type":"ContainerStarted","Data":"9b513ecfadfbd0765ffaaa5bf1a43810790553ea796377c81b4deff0a9555415"} Dec 06 04:07:07 crc kubenswrapper[4980]: I1206 04:07:07.850686 4980 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83brxgn_73b1ef5a-140c-40df-a801-dc06cc56b5e2/util/0.log" Dec 06 04:07:07 crc kubenswrapper[4980]: I1206 04:07:07.891192 4980 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83brxgn_73b1ef5a-140c-40df-a801-dc06cc56b5e2/pull/0.log" Dec 06 04:07:07 crc kubenswrapper[4980]: I1206 04:07:07.893340 4980 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83brxgn_73b1ef5a-140c-40df-a801-dc06cc56b5e2/pull/0.log" Dec 06 04:07:07 crc kubenswrapper[4980]: I1206 04:07:07.966305 4980 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83brxgn_73b1ef5a-140c-40df-a801-dc06cc56b5e2/util/0.log" Dec 06 04:07:08 crc kubenswrapper[4980]: I1206 04:07:08.090389 4980 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83brxgn_73b1ef5a-140c-40df-a801-dc06cc56b5e2/pull/0.log" Dec 06 04:07:08 crc kubenswrapper[4980]: I1206 04:07:08.201474 4980 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83brxgn_73b1ef5a-140c-40df-a801-dc06cc56b5e2/extract/0.log" Dec 06 04:07:08 crc kubenswrapper[4980]: I1206 04:07:08.246610 4980 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-d2hwb_6b108c98-9b6c-48f5-a6a9-2db36a4e5283/extract-utilities/0.log" Dec 06 04:07:08 crc kubenswrapper[4980]: I1206 04:07:08.545876 4980 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-d2hwb_6b108c98-9b6c-48f5-a6a9-2db36a4e5283/extract-content/0.log" Dec 06 04:07:08 crc kubenswrapper[4980]: I1206 04:07:08.567976 4980 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-d2hwb_6b108c98-9b6c-48f5-a6a9-2db36a4e5283/extract-content/0.log" Dec 06 04:07:08 crc kubenswrapper[4980]: I1206 04:07:08.574663 4980 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-d2hwb_6b108c98-9b6c-48f5-a6a9-2db36a4e5283/extract-utilities/0.log" Dec 06 04:07:08 crc kubenswrapper[4980]: I1206 04:07:08.722205 4980 generic.go:334] "Generic (PLEG): container finished" podID="cd6724f3-6ebc-4c26-9d2c-eaa0de6c503c" containerID="9be0ebeb1c8449ef2df16ff1fb33c569bbd4f55a6635378bcc0813b0963632b2" exitCode=0 Dec 06 04:07:08 crc kubenswrapper[4980]: I1206 04:07:08.722250 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-tgs44" event={"ID":"cd6724f3-6ebc-4c26-9d2c-eaa0de6c503c","Type":"ContainerDied","Data":"9be0ebeb1c8449ef2df16ff1fb33c569bbd4f55a6635378bcc0813b0963632b2"} Dec 06 04:07:08 crc kubenswrapper[4980]: I1206 04:07:08.724101 4980 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 06 04:07:08 crc kubenswrapper[4980]: I1206 04:07:08.834743 4980 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-d2hwb_6b108c98-9b6c-48f5-a6a9-2db36a4e5283/extract-content/0.log" Dec 06 04:07:08 crc kubenswrapper[4980]: I1206 04:07:08.851767 4980 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-d2hwb_6b108c98-9b6c-48f5-a6a9-2db36a4e5283/extract-utilities/0.log" Dec 06 04:07:09 crc kubenswrapper[4980]: I1206 04:07:09.067430 4980 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-x8x8b_b6205a6a-fc2b-4a1e-aefd-b86867f3478d/extract-utilities/0.log" Dec 06 04:07:09 crc kubenswrapper[4980]: I1206 04:07:09.235947 4980 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-d2hwb_6b108c98-9b6c-48f5-a6a9-2db36a4e5283/registry-server/0.log" Dec 06 04:07:09 crc kubenswrapper[4980]: I1206 04:07:09.236281 4980 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-x8x8b_b6205a6a-fc2b-4a1e-aefd-b86867f3478d/extract-utilities/0.log" Dec 06 04:07:09 crc kubenswrapper[4980]: I1206 04:07:09.260661 4980 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-x8x8b_b6205a6a-fc2b-4a1e-aefd-b86867f3478d/extract-content/0.log" Dec 06 04:07:09 crc kubenswrapper[4980]: I1206 04:07:09.295592 4980 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-x8x8b_b6205a6a-fc2b-4a1e-aefd-b86867f3478d/extract-content/0.log" Dec 06 04:07:09 crc kubenswrapper[4980]: I1206 04:07:09.559488 4980 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-x8x8b_b6205a6a-fc2b-4a1e-aefd-b86867f3478d/extract-utilities/0.log" Dec 06 04:07:09 crc kubenswrapper[4980]: I1206 04:07:09.578190 4980 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-x8x8b_b6205a6a-fc2b-4a1e-aefd-b86867f3478d/extract-content/0.log" Dec 06 04:07:09 crc kubenswrapper[4980]: I1206 04:07:09.732818 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-tgs44" event={"ID":"cd6724f3-6ebc-4c26-9d2c-eaa0de6c503c","Type":"ContainerStarted","Data":"5fc761b67ce794d97aedbb15a28c08071dcec71888b378d741b4ff72b5093d9f"} Dec 06 04:07:09 crc kubenswrapper[4980]: I1206 04:07:09.848168 4980 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-tgld8_2392c238-d47c-472f-b8f3-cf6ad2720454/extract-utilities/0.log" Dec 06 04:07:09 crc kubenswrapper[4980]: I1206 04:07:09.919545 4980 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-rhhmp_34c9fd26-4164-4267-8dce-7f487139e741/marketplace-operator/0.log" Dec 06 04:07:09 crc kubenswrapper[4980]: I1206 04:07:09.977874 4980 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-x8x8b_b6205a6a-fc2b-4a1e-aefd-b86867f3478d/registry-server/0.log" Dec 06 04:07:10 crc kubenswrapper[4980]: I1206 04:07:10.053184 4980 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-tgld8_2392c238-d47c-472f-b8f3-cf6ad2720454/extract-content/0.log" Dec 06 04:07:10 crc kubenswrapper[4980]: I1206 04:07:10.103134 4980 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-tgld8_2392c238-d47c-472f-b8f3-cf6ad2720454/extract-utilities/0.log" Dec 06 04:07:10 crc kubenswrapper[4980]: I1206 04:07:10.125378 4980 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-tgld8_2392c238-d47c-472f-b8f3-cf6ad2720454/extract-content/0.log" Dec 06 04:07:10 crc kubenswrapper[4980]: I1206 04:07:10.342033 4980 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-tgld8_2392c238-d47c-472f-b8f3-cf6ad2720454/extract-utilities/0.log" Dec 06 04:07:10 crc kubenswrapper[4980]: I1206 04:07:10.374916 4980 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-tgld8_2392c238-d47c-472f-b8f3-cf6ad2720454/extract-content/0.log" Dec 06 04:07:10 crc kubenswrapper[4980]: I1206 04:07:10.457311 4980 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-tgld8_2392c238-d47c-472f-b8f3-cf6ad2720454/registry-server/0.log" Dec 06 04:07:10 crc kubenswrapper[4980]: I1206 04:07:10.577772 4980 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-kfc6d_aca9683f-a0d4-46f8-aacb-f2e45896458c/extract-utilities/0.log" Dec 06 04:07:10 crc kubenswrapper[4980]: I1206 04:07:10.708892 4980 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-kfc6d_aca9683f-a0d4-46f8-aacb-f2e45896458c/extract-content/0.log" Dec 06 04:07:10 crc kubenswrapper[4980]: I1206 04:07:10.741884 4980 generic.go:334] "Generic (PLEG): container finished" podID="cd6724f3-6ebc-4c26-9d2c-eaa0de6c503c" containerID="5fc761b67ce794d97aedbb15a28c08071dcec71888b378d741b4ff72b5093d9f" exitCode=0 Dec 06 04:07:10 crc kubenswrapper[4980]: I1206 04:07:10.741933 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-tgs44" event={"ID":"cd6724f3-6ebc-4c26-9d2c-eaa0de6c503c","Type":"ContainerDied","Data":"5fc761b67ce794d97aedbb15a28c08071dcec71888b378d741b4ff72b5093d9f"} Dec 06 04:07:10 crc kubenswrapper[4980]: I1206 04:07:10.755097 4980 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-kfc6d_aca9683f-a0d4-46f8-aacb-f2e45896458c/extract-utilities/0.log" Dec 06 04:07:10 crc kubenswrapper[4980]: I1206 04:07:10.765670 4980 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-kfc6d_aca9683f-a0d4-46f8-aacb-f2e45896458c/extract-content/0.log" Dec 06 04:07:10 crc kubenswrapper[4980]: I1206 04:07:10.946971 4980 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-kfc6d_aca9683f-a0d4-46f8-aacb-f2e45896458c/extract-content/0.log" Dec 06 04:07:10 crc kubenswrapper[4980]: I1206 04:07:10.982460 4980 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-kfc6d_aca9683f-a0d4-46f8-aacb-f2e45896458c/extract-utilities/0.log" Dec 06 04:07:11 crc kubenswrapper[4980]: I1206 04:07:11.207504 4980 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-tgs44_cd6724f3-6ebc-4c26-9d2c-eaa0de6c503c/extract-utilities/0.log" Dec 06 04:07:11 crc kubenswrapper[4980]: I1206 04:07:11.320390 4980 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-kfc6d_aca9683f-a0d4-46f8-aacb-f2e45896458c/registry-server/0.log" Dec 06 04:07:11 crc kubenswrapper[4980]: I1206 04:07:11.356031 4980 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-tgs44_cd6724f3-6ebc-4c26-9d2c-eaa0de6c503c/extract-utilities/0.log" Dec 06 04:07:11 crc kubenswrapper[4980]: I1206 04:07:11.368959 4980 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-tgs44_cd6724f3-6ebc-4c26-9d2c-eaa0de6c503c/extract-content/0.log" Dec 06 04:07:11 crc kubenswrapper[4980]: I1206 04:07:11.403350 4980 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-tgs44_cd6724f3-6ebc-4c26-9d2c-eaa0de6c503c/extract-content/0.log" Dec 06 04:07:11 crc kubenswrapper[4980]: I1206 04:07:11.711793 4980 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-tgs44_cd6724f3-6ebc-4c26-9d2c-eaa0de6c503c/extract-utilities/0.log" Dec 06 04:07:11 crc kubenswrapper[4980]: I1206 04:07:11.743893 4980 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-tgs44_cd6724f3-6ebc-4c26-9d2c-eaa0de6c503c/extract-content/0.log" Dec 06 04:07:11 crc kubenswrapper[4980]: I1206 04:07:11.752915 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-tgs44" event={"ID":"cd6724f3-6ebc-4c26-9d2c-eaa0de6c503c","Type":"ContainerStarted","Data":"192197a6cdb5faade83ae747567dae84915735a4acdb08713c5e75ffcfc68c98"} Dec 06 04:07:11 crc kubenswrapper[4980]: I1206 04:07:11.775289 4980 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-tgs44" podStartSLOduration=3.29595591 podStartE2EDuration="5.775251701s" podCreationTimestamp="2025-12-06 04:07:06 +0000 UTC" firstStartedPulling="2025-12-06 04:07:08.723834271 +0000 UTC m=+1987.968585542" lastFinishedPulling="2025-12-06 04:07:11.203130062 +0000 UTC m=+1990.447881333" observedRunningTime="2025-12-06 04:07:11.774108989 +0000 UTC m=+1991.018860270" watchObservedRunningTime="2025-12-06 04:07:11.775251701 +0000 UTC m=+1991.020002972" Dec 06 04:07:13 crc kubenswrapper[4980]: I1206 04:07:13.037334 4980 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-db-create-4wjqb"] Dec 06 04:07:13 crc kubenswrapper[4980]: I1206 04:07:13.045169 4980 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-db-create-4wjqb"] Dec 06 04:07:13 crc kubenswrapper[4980]: I1206 04:07:13.203057 4980 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d9f133d0-65f5-4ee3-9906-97e1a883ddf8" path="/var/lib/kubelet/pods/d9f133d0-65f5-4ee3-9906-97e1a883ddf8/volumes" Dec 06 04:07:17 crc kubenswrapper[4980]: I1206 04:07:17.106523 4980 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-tgs44" Dec 06 04:07:17 crc kubenswrapper[4980]: I1206 04:07:17.106953 4980 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-tgs44" Dec 06 04:07:17 crc kubenswrapper[4980]: I1206 04:07:17.157745 4980 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-tgs44" Dec 06 04:07:17 crc kubenswrapper[4980]: I1206 04:07:17.942233 4980 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-tgs44" Dec 06 04:07:17 crc kubenswrapper[4980]: I1206 04:07:17.995311 4980 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-tgs44"] Dec 06 04:07:19 crc kubenswrapper[4980]: I1206 04:07:19.897495 4980 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-tgs44" podUID="cd6724f3-6ebc-4c26-9d2c-eaa0de6c503c" containerName="registry-server" containerID="cri-o://192197a6cdb5faade83ae747567dae84915735a4acdb08713c5e75ffcfc68c98" gracePeriod=2 Dec 06 04:07:21 crc kubenswrapper[4980]: I1206 04:07:21.916485 4980 generic.go:334] "Generic (PLEG): container finished" podID="cd6724f3-6ebc-4c26-9d2c-eaa0de6c503c" containerID="192197a6cdb5faade83ae747567dae84915735a4acdb08713c5e75ffcfc68c98" exitCode=0 Dec 06 04:07:21 crc kubenswrapper[4980]: I1206 04:07:21.916548 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-tgs44" event={"ID":"cd6724f3-6ebc-4c26-9d2c-eaa0de6c503c","Type":"ContainerDied","Data":"192197a6cdb5faade83ae747567dae84915735a4acdb08713c5e75ffcfc68c98"} Dec 06 04:07:22 crc kubenswrapper[4980]: I1206 04:07:22.028911 4980 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-8c70-account-create-mkrvh"] Dec 06 04:07:22 crc kubenswrapper[4980]: I1206 04:07:22.042243 4980 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-8c70-account-create-mkrvh"] Dec 06 04:07:22 crc kubenswrapper[4980]: I1206 04:07:22.500755 4980 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-tgs44" Dec 06 04:07:22 crc kubenswrapper[4980]: I1206 04:07:22.666159 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cd6724f3-6ebc-4c26-9d2c-eaa0de6c503c-utilities\") pod \"cd6724f3-6ebc-4c26-9d2c-eaa0de6c503c\" (UID: \"cd6724f3-6ebc-4c26-9d2c-eaa0de6c503c\") " Dec 06 04:07:22 crc kubenswrapper[4980]: I1206 04:07:22.666336 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qf897\" (UniqueName: \"kubernetes.io/projected/cd6724f3-6ebc-4c26-9d2c-eaa0de6c503c-kube-api-access-qf897\") pod \"cd6724f3-6ebc-4c26-9d2c-eaa0de6c503c\" (UID: \"cd6724f3-6ebc-4c26-9d2c-eaa0de6c503c\") " Dec 06 04:07:22 crc kubenswrapper[4980]: I1206 04:07:22.666405 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cd6724f3-6ebc-4c26-9d2c-eaa0de6c503c-catalog-content\") pod \"cd6724f3-6ebc-4c26-9d2c-eaa0de6c503c\" (UID: \"cd6724f3-6ebc-4c26-9d2c-eaa0de6c503c\") " Dec 06 04:07:22 crc kubenswrapper[4980]: I1206 04:07:22.667487 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cd6724f3-6ebc-4c26-9d2c-eaa0de6c503c-utilities" (OuterVolumeSpecName: "utilities") pod "cd6724f3-6ebc-4c26-9d2c-eaa0de6c503c" (UID: "cd6724f3-6ebc-4c26-9d2c-eaa0de6c503c"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 04:07:22 crc kubenswrapper[4980]: I1206 04:07:22.675721 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cd6724f3-6ebc-4c26-9d2c-eaa0de6c503c-kube-api-access-qf897" (OuterVolumeSpecName: "kube-api-access-qf897") pod "cd6724f3-6ebc-4c26-9d2c-eaa0de6c503c" (UID: "cd6724f3-6ebc-4c26-9d2c-eaa0de6c503c"). InnerVolumeSpecName "kube-api-access-qf897". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 04:07:22 crc kubenswrapper[4980]: I1206 04:07:22.768973 4980 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cd6724f3-6ebc-4c26-9d2c-eaa0de6c503c-utilities\") on node \"crc\" DevicePath \"\"" Dec 06 04:07:22 crc kubenswrapper[4980]: I1206 04:07:22.769026 4980 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qf897\" (UniqueName: \"kubernetes.io/projected/cd6724f3-6ebc-4c26-9d2c-eaa0de6c503c-kube-api-access-qf897\") on node \"crc\" DevicePath \"\"" Dec 06 04:07:22 crc kubenswrapper[4980]: I1206 04:07:22.846047 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cd6724f3-6ebc-4c26-9d2c-eaa0de6c503c-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "cd6724f3-6ebc-4c26-9d2c-eaa0de6c503c" (UID: "cd6724f3-6ebc-4c26-9d2c-eaa0de6c503c"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 04:07:22 crc kubenswrapper[4980]: I1206 04:07:22.869996 4980 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cd6724f3-6ebc-4c26-9d2c-eaa0de6c503c-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 06 04:07:22 crc kubenswrapper[4980]: I1206 04:07:22.926016 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-tgs44" event={"ID":"cd6724f3-6ebc-4c26-9d2c-eaa0de6c503c","Type":"ContainerDied","Data":"9b513ecfadfbd0765ffaaa5bf1a43810790553ea796377c81b4deff0a9555415"} Dec 06 04:07:22 crc kubenswrapper[4980]: I1206 04:07:22.926099 4980 scope.go:117] "RemoveContainer" containerID="192197a6cdb5faade83ae747567dae84915735a4acdb08713c5e75ffcfc68c98" Dec 06 04:07:22 crc kubenswrapper[4980]: I1206 04:07:22.926241 4980 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-tgs44" Dec 06 04:07:22 crc kubenswrapper[4980]: I1206 04:07:22.957849 4980 scope.go:117] "RemoveContainer" containerID="5fc761b67ce794d97aedbb15a28c08071dcec71888b378d741b4ff72b5093d9f" Dec 06 04:07:22 crc kubenswrapper[4980]: I1206 04:07:22.974470 4980 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-tgs44"] Dec 06 04:07:22 crc kubenswrapper[4980]: I1206 04:07:22.983669 4980 scope.go:117] "RemoveContainer" containerID="9be0ebeb1c8449ef2df16ff1fb33c569bbd4f55a6635378bcc0813b0963632b2" Dec 06 04:07:22 crc kubenswrapper[4980]: I1206 04:07:22.989828 4980 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-tgs44"] Dec 06 04:07:23 crc kubenswrapper[4980]: E1206 04:07:23.056592 4980 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podcd6724f3_6ebc_4c26_9d2c_eaa0de6c503c.slice\": RecentStats: unable to find data in memory cache]" Dec 06 04:07:23 crc kubenswrapper[4980]: I1206 04:07:23.196534 4980 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="aae1a12f-6f92-43cc-99ae-fd14a894aadd" path="/var/lib/kubelet/pods/aae1a12f-6f92-43cc-99ae-fd14a894aadd/volumes" Dec 06 04:07:23 crc kubenswrapper[4980]: I1206 04:07:23.197644 4980 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cd6724f3-6ebc-4c26-9d2c-eaa0de6c503c" path="/var/lib/kubelet/pods/cd6724f3-6ebc-4c26-9d2c-eaa0de6c503c/volumes" Dec 06 04:07:25 crc kubenswrapper[4980]: I1206 04:07:25.137586 4980 patch_prober.go:28] interesting pod/machine-config-daemon-r5zfb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 04:07:25 crc kubenswrapper[4980]: I1206 04:07:25.137917 4980 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-r5zfb" podUID="320f44d1-a671-4a91-b328-a8b0fdd8f23a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 04:07:29 crc kubenswrapper[4980]: I1206 04:07:29.041546 4980 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-db-sync-9lbg6"] Dec 06 04:07:29 crc kubenswrapper[4980]: I1206 04:07:29.047052 4980 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-db-sync-9lbg6"] Dec 06 04:07:29 crc kubenswrapper[4980]: I1206 04:07:29.195770 4980 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="34f14e45-81a8-490f-a760-88fbac6f06b3" path="/var/lib/kubelet/pods/34f14e45-81a8-490f-a760-88fbac6f06b3/volumes" Dec 06 04:07:55 crc kubenswrapper[4980]: I1206 04:07:55.138098 4980 patch_prober.go:28] interesting pod/machine-config-daemon-r5zfb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 04:07:55 crc kubenswrapper[4980]: I1206 04:07:55.138794 4980 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-r5zfb" podUID="320f44d1-a671-4a91-b328-a8b0fdd8f23a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 04:07:55 crc kubenswrapper[4980]: I1206 04:07:55.138864 4980 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-r5zfb" Dec 06 04:07:55 crc kubenswrapper[4980]: I1206 04:07:55.139632 4980 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"7e5a814e12c0fd9a500ff15c5dc09dfe0e3f68d5cc913a88b3e14e53b9b65070"} pod="openshift-machine-config-operator/machine-config-daemon-r5zfb" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 06 04:07:55 crc kubenswrapper[4980]: I1206 04:07:55.139732 4980 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-r5zfb" podUID="320f44d1-a671-4a91-b328-a8b0fdd8f23a" containerName="machine-config-daemon" containerID="cri-o://7e5a814e12c0fd9a500ff15c5dc09dfe0e3f68d5cc913a88b3e14e53b9b65070" gracePeriod=600 Dec 06 04:07:56 crc kubenswrapper[4980]: I1206 04:07:56.285730 4980 generic.go:334] "Generic (PLEG): container finished" podID="320f44d1-a671-4a91-b328-a8b0fdd8f23a" containerID="7e5a814e12c0fd9a500ff15c5dc09dfe0e3f68d5cc913a88b3e14e53b9b65070" exitCode=0 Dec 06 04:07:56 crc kubenswrapper[4980]: I1206 04:07:56.286266 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-r5zfb" event={"ID":"320f44d1-a671-4a91-b328-a8b0fdd8f23a","Type":"ContainerDied","Data":"7e5a814e12c0fd9a500ff15c5dc09dfe0e3f68d5cc913a88b3e14e53b9b65070"} Dec 06 04:07:56 crc kubenswrapper[4980]: I1206 04:07:56.286704 4980 scope.go:117] "RemoveContainer" containerID="e8878e13c398cceec46bb52c4c9ea0e06dbf4317e6ad39b69ca763a0d8b15497" Dec 06 04:07:56 crc kubenswrapper[4980]: I1206 04:07:56.286305 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-r5zfb" event={"ID":"320f44d1-a671-4a91-b328-a8b0fdd8f23a","Type":"ContainerStarted","Data":"dfbce6b7cffe7859167e3130a4ee8cae3ec15169d967291de42294eb241ffd7a"} Dec 06 04:08:07 crc kubenswrapper[4980]: I1206 04:08:07.006437 4980 scope.go:117] "RemoveContainer" containerID="4b5687718ac0054924b11fbba5a67c27ce606bd67f575fd23dfa13402b408f11" Dec 06 04:08:07 crc kubenswrapper[4980]: I1206 04:08:07.046634 4980 scope.go:117] "RemoveContainer" containerID="31ffa96d7d8a317ed02c26ebf3a14d8114b58fb0da24d85117a2c65a3d8b80c7" Dec 06 04:08:07 crc kubenswrapper[4980]: I1206 04:08:07.137811 4980 scope.go:117] "RemoveContainer" containerID="be215c655f9b5a0857ea6486ec2451c3717788171774e8b9f09a21787b8136d9" Dec 06 04:08:19 crc kubenswrapper[4980]: I1206 04:08:19.554605 4980 generic.go:334] "Generic (PLEG): container finished" podID="35cb639f-8637-4bd9-906a-d83988d015ff" containerID="8c0c956e04fbd4dad778a309ea2058c2348a4e65773bf8bb813e89a67e4fe405" exitCode=0 Dec 06 04:08:19 crc kubenswrapper[4980]: I1206 04:08:19.554728 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-bgmb7/must-gather-d8t8w" event={"ID":"35cb639f-8637-4bd9-906a-d83988d015ff","Type":"ContainerDied","Data":"8c0c956e04fbd4dad778a309ea2058c2348a4e65773bf8bb813e89a67e4fe405"} Dec 06 04:08:19 crc kubenswrapper[4980]: I1206 04:08:19.555768 4980 scope.go:117] "RemoveContainer" containerID="8c0c956e04fbd4dad778a309ea2058c2348a4e65773bf8bb813e89a67e4fe405" Dec 06 04:08:20 crc kubenswrapper[4980]: I1206 04:08:20.105466 4980 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-bgmb7_must-gather-d8t8w_35cb639f-8637-4bd9-906a-d83988d015ff/gather/0.log" Dec 06 04:08:27 crc kubenswrapper[4980]: I1206 04:08:27.128717 4980 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-bgmb7/must-gather-d8t8w"] Dec 06 04:08:27 crc kubenswrapper[4980]: I1206 04:08:27.129578 4980 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-must-gather-bgmb7/must-gather-d8t8w" podUID="35cb639f-8637-4bd9-906a-d83988d015ff" containerName="copy" containerID="cri-o://cf7175d2595ea07c5997befacc560760e8ab20ff7b7e2a469abcaad88333d487" gracePeriod=2 Dec 06 04:08:27 crc kubenswrapper[4980]: I1206 04:08:27.136122 4980 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-bgmb7/must-gather-d8t8w"] Dec 06 04:08:27 crc kubenswrapper[4980]: I1206 04:08:27.531916 4980 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-bgmb7_must-gather-d8t8w_35cb639f-8637-4bd9-906a-d83988d015ff/copy/0.log" Dec 06 04:08:27 crc kubenswrapper[4980]: I1206 04:08:27.532746 4980 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-bgmb7/must-gather-d8t8w" Dec 06 04:08:27 crc kubenswrapper[4980]: I1206 04:08:27.636422 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/35cb639f-8637-4bd9-906a-d83988d015ff-must-gather-output\") pod \"35cb639f-8637-4bd9-906a-d83988d015ff\" (UID: \"35cb639f-8637-4bd9-906a-d83988d015ff\") " Dec 06 04:08:27 crc kubenswrapper[4980]: I1206 04:08:27.636485 4980 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tpr6q\" (UniqueName: \"kubernetes.io/projected/35cb639f-8637-4bd9-906a-d83988d015ff-kube-api-access-tpr6q\") pod \"35cb639f-8637-4bd9-906a-d83988d015ff\" (UID: \"35cb639f-8637-4bd9-906a-d83988d015ff\") " Dec 06 04:08:27 crc kubenswrapper[4980]: I1206 04:08:27.641758 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/35cb639f-8637-4bd9-906a-d83988d015ff-kube-api-access-tpr6q" (OuterVolumeSpecName: "kube-api-access-tpr6q") pod "35cb639f-8637-4bd9-906a-d83988d015ff" (UID: "35cb639f-8637-4bd9-906a-d83988d015ff"). InnerVolumeSpecName "kube-api-access-tpr6q". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 04:08:27 crc kubenswrapper[4980]: I1206 04:08:27.696779 4980 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-bgmb7_must-gather-d8t8w_35cb639f-8637-4bd9-906a-d83988d015ff/copy/0.log" Dec 06 04:08:27 crc kubenswrapper[4980]: I1206 04:08:27.697583 4980 generic.go:334] "Generic (PLEG): container finished" podID="35cb639f-8637-4bd9-906a-d83988d015ff" containerID="cf7175d2595ea07c5997befacc560760e8ab20ff7b7e2a469abcaad88333d487" exitCode=143 Dec 06 04:08:27 crc kubenswrapper[4980]: I1206 04:08:27.697638 4980 scope.go:117] "RemoveContainer" containerID="cf7175d2595ea07c5997befacc560760e8ab20ff7b7e2a469abcaad88333d487" Dec 06 04:08:27 crc kubenswrapper[4980]: I1206 04:08:27.697674 4980 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-bgmb7/must-gather-d8t8w" Dec 06 04:08:27 crc kubenswrapper[4980]: I1206 04:08:27.717626 4980 scope.go:117] "RemoveContainer" containerID="8c0c956e04fbd4dad778a309ea2058c2348a4e65773bf8bb813e89a67e4fe405" Dec 06 04:08:27 crc kubenswrapper[4980]: I1206 04:08:27.729531 4980 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/35cb639f-8637-4bd9-906a-d83988d015ff-must-gather-output" (OuterVolumeSpecName: "must-gather-output") pod "35cb639f-8637-4bd9-906a-d83988d015ff" (UID: "35cb639f-8637-4bd9-906a-d83988d015ff"). InnerVolumeSpecName "must-gather-output". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 04:08:27 crc kubenswrapper[4980]: I1206 04:08:27.738172 4980 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tpr6q\" (UniqueName: \"kubernetes.io/projected/35cb639f-8637-4bd9-906a-d83988d015ff-kube-api-access-tpr6q\") on node \"crc\" DevicePath \"\"" Dec 06 04:08:27 crc kubenswrapper[4980]: I1206 04:08:27.738198 4980 reconciler_common.go:293] "Volume detached for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/35cb639f-8637-4bd9-906a-d83988d015ff-must-gather-output\") on node \"crc\" DevicePath \"\"" Dec 06 04:08:27 crc kubenswrapper[4980]: I1206 04:08:27.761609 4980 scope.go:117] "RemoveContainer" containerID="cf7175d2595ea07c5997befacc560760e8ab20ff7b7e2a469abcaad88333d487" Dec 06 04:08:27 crc kubenswrapper[4980]: E1206 04:08:27.762151 4980 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cf7175d2595ea07c5997befacc560760e8ab20ff7b7e2a469abcaad88333d487\": container with ID starting with cf7175d2595ea07c5997befacc560760e8ab20ff7b7e2a469abcaad88333d487 not found: ID does not exist" containerID="cf7175d2595ea07c5997befacc560760e8ab20ff7b7e2a469abcaad88333d487" Dec 06 04:08:27 crc kubenswrapper[4980]: I1206 04:08:27.762197 4980 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cf7175d2595ea07c5997befacc560760e8ab20ff7b7e2a469abcaad88333d487"} err="failed to get container status \"cf7175d2595ea07c5997befacc560760e8ab20ff7b7e2a469abcaad88333d487\": rpc error: code = NotFound desc = could not find container \"cf7175d2595ea07c5997befacc560760e8ab20ff7b7e2a469abcaad88333d487\": container with ID starting with cf7175d2595ea07c5997befacc560760e8ab20ff7b7e2a469abcaad88333d487 not found: ID does not exist" Dec 06 04:08:27 crc kubenswrapper[4980]: I1206 04:08:27.762228 4980 scope.go:117] "RemoveContainer" containerID="8c0c956e04fbd4dad778a309ea2058c2348a4e65773bf8bb813e89a67e4fe405" Dec 06 04:08:27 crc kubenswrapper[4980]: E1206 04:08:27.762540 4980 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8c0c956e04fbd4dad778a309ea2058c2348a4e65773bf8bb813e89a67e4fe405\": container with ID starting with 8c0c956e04fbd4dad778a309ea2058c2348a4e65773bf8bb813e89a67e4fe405 not found: ID does not exist" containerID="8c0c956e04fbd4dad778a309ea2058c2348a4e65773bf8bb813e89a67e4fe405" Dec 06 04:08:27 crc kubenswrapper[4980]: I1206 04:08:27.762630 4980 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8c0c956e04fbd4dad778a309ea2058c2348a4e65773bf8bb813e89a67e4fe405"} err="failed to get container status \"8c0c956e04fbd4dad778a309ea2058c2348a4e65773bf8bb813e89a67e4fe405\": rpc error: code = NotFound desc = could not find container \"8c0c956e04fbd4dad778a309ea2058c2348a4e65773bf8bb813e89a67e4fe405\": container with ID starting with 8c0c956e04fbd4dad778a309ea2058c2348a4e65773bf8bb813e89a67e4fe405 not found: ID does not exist" Dec 06 04:08:29 crc kubenswrapper[4980]: I1206 04:08:29.230579 4980 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="35cb639f-8637-4bd9-906a-d83988d015ff" path="/var/lib/kubelet/pods/35cb639f-8637-4bd9-906a-d83988d015ff/volumes" Dec 06 04:09:55 crc kubenswrapper[4980]: I1206 04:09:55.137917 4980 patch_prober.go:28] interesting pod/machine-config-daemon-r5zfb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 04:09:55 crc kubenswrapper[4980]: I1206 04:09:55.138973 4980 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-r5zfb" podUID="320f44d1-a671-4a91-b328-a8b0fdd8f23a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 04:10:25 crc kubenswrapper[4980]: I1206 04:10:25.137452 4980 patch_prober.go:28] interesting pod/machine-config-daemon-r5zfb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 04:10:25 crc kubenswrapper[4980]: I1206 04:10:25.138376 4980 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-r5zfb" podUID="320f44d1-a671-4a91-b328-a8b0fdd8f23a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 04:10:55 crc kubenswrapper[4980]: I1206 04:10:55.138318 4980 patch_prober.go:28] interesting pod/machine-config-daemon-r5zfb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 04:10:55 crc kubenswrapper[4980]: I1206 04:10:55.139010 4980 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-r5zfb" podUID="320f44d1-a671-4a91-b328-a8b0fdd8f23a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 04:10:55 crc kubenswrapper[4980]: I1206 04:10:55.139067 4980 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-r5zfb" Dec 06 04:10:55 crc kubenswrapper[4980]: I1206 04:10:55.139854 4980 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"dfbce6b7cffe7859167e3130a4ee8cae3ec15169d967291de42294eb241ffd7a"} pod="openshift-machine-config-operator/machine-config-daemon-r5zfb" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 06 04:10:55 crc kubenswrapper[4980]: I1206 04:10:55.139967 4980 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-r5zfb" podUID="320f44d1-a671-4a91-b328-a8b0fdd8f23a" containerName="machine-config-daemon" containerID="cri-o://dfbce6b7cffe7859167e3130a4ee8cae3ec15169d967291de42294eb241ffd7a" gracePeriod=600 Dec 06 04:10:55 crc kubenswrapper[4980]: E1206 04:10:55.268037 4980 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-r5zfb_openshift-machine-config-operator(320f44d1-a671-4a91-b328-a8b0fdd8f23a)\"" pod="openshift-machine-config-operator/machine-config-daemon-r5zfb" podUID="320f44d1-a671-4a91-b328-a8b0fdd8f23a" Dec 06 04:10:56 crc kubenswrapper[4980]: I1206 04:10:56.248270 4980 generic.go:334] "Generic (PLEG): container finished" podID="320f44d1-a671-4a91-b328-a8b0fdd8f23a" containerID="dfbce6b7cffe7859167e3130a4ee8cae3ec15169d967291de42294eb241ffd7a" exitCode=0 Dec 06 04:10:56 crc kubenswrapper[4980]: I1206 04:10:56.248342 4980 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-r5zfb" event={"ID":"320f44d1-a671-4a91-b328-a8b0fdd8f23a","Type":"ContainerDied","Data":"dfbce6b7cffe7859167e3130a4ee8cae3ec15169d967291de42294eb241ffd7a"} Dec 06 04:10:56 crc kubenswrapper[4980]: I1206 04:10:56.248716 4980 scope.go:117] "RemoveContainer" containerID="7e5a814e12c0fd9a500ff15c5dc09dfe0e3f68d5cc913a88b3e14e53b9b65070" Dec 06 04:10:56 crc kubenswrapper[4980]: I1206 04:10:56.249830 4980 scope.go:117] "RemoveContainer" containerID="dfbce6b7cffe7859167e3130a4ee8cae3ec15169d967291de42294eb241ffd7a" Dec 06 04:10:56 crc kubenswrapper[4980]: E1206 04:10:56.250189 4980 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-r5zfb_openshift-machine-config-operator(320f44d1-a671-4a91-b328-a8b0fdd8f23a)\"" pod="openshift-machine-config-operator/machine-config-daemon-r5zfb" podUID="320f44d1-a671-4a91-b328-a8b0fdd8f23a" Dec 06 04:11:08 crc kubenswrapper[4980]: I1206 04:11:08.184623 4980 scope.go:117] "RemoveContainer" containerID="dfbce6b7cffe7859167e3130a4ee8cae3ec15169d967291de42294eb241ffd7a" Dec 06 04:11:08 crc kubenswrapper[4980]: E1206 04:11:08.185531 4980 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-r5zfb_openshift-machine-config-operator(320f44d1-a671-4a91-b328-a8b0fdd8f23a)\"" pod="openshift-machine-config-operator/machine-config-daemon-r5zfb" podUID="320f44d1-a671-4a91-b328-a8b0fdd8f23a" Dec 06 04:11:19 crc kubenswrapper[4980]: I1206 04:11:19.184540 4980 scope.go:117] "RemoveContainer" containerID="dfbce6b7cffe7859167e3130a4ee8cae3ec15169d967291de42294eb241ffd7a" Dec 06 04:11:19 crc kubenswrapper[4980]: E1206 04:11:19.185311 4980 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-r5zfb_openshift-machine-config-operator(320f44d1-a671-4a91-b328-a8b0fdd8f23a)\"" pod="openshift-machine-config-operator/machine-config-daemon-r5zfb" podUID="320f44d1-a671-4a91-b328-a8b0fdd8f23a" var/home/core/zuul-output/logs/crc-cloud-workdir-crc-all-logs.tar.gz0000644000175000000000000000005515114726365024457 0ustar coreroot‹íÁ  ÷Om7 €7šÞ'(var/home/core/zuul-output/logs/crc-cloud/0000755000175000000000000000000015114726366017375 5ustar corerootvar/home/core/zuul-output/artifacts/0000755000175000017500000000000015114721600016502 5ustar corecorevar/home/core/zuul-output/docs/0000755000175000017500000000000015114721600015452 5ustar corecore